The Project Physics Course
Reader
Light and Electromagnetism
The Project Physics Course
Reader
UNIT
4 Light and Electromagnetism
A Component of the
Project Physics Course
Published by
HOLT, RINEHART and WINSTON, Inc.
New York, Toronto
This publication is one of the many
instructional materials developed for the
Project Physics Course. These materials
include Texts, Handbooks, Teacher Resource
Books, Readers, Programmed Instruction
Booklets, Film Loops, Transparencies, 16mm
films and laboratory equipment. Development
of the course has profited from the help of
many colleagues listed in the text units.
Directors of Harvard Project Physics
Gerald Holton, Department of Physics,
Harvard University
F. James Rutherford, Capuchino High School,
San Bruno, California, and Harvard University
Fletcher G. Watson, Harvard Graduate School
of Education
Copyright © 1 971 , Project Physics
All Rights Reserved
SBN 03-084561-0
1234 039 98765432
Project Physics is a registered trademark
Picture Credits
Cover: Current, 1964, by Bridget Riley. Emulsion on
composition board, 583/8 x 587/8". Courtesy of
The Museum of Modern Art, New York City.
2 «
5 I
3 6
Picture Credits for frontispiece.
(1) Photograph by Glen J. Pearcy.
(2) Jeune fille au corsage rouge lisant by Jean
Baptiste Camille Corot. Painting. Collection
Biihrle, Zurich.
(3) Harvard Project Physics staff photo.
(4) Femme lisant by Georges Seurat. Conte crayon
drawing. Collection C. F. Stooo. London.
(5) Portrait of Pierre Reverdy by Pablo Picasso.
Etching. Museum of Modern Art, N.Y.C.
(6) Lecture au lit by Paul Klee. Drawing. Paul Klee
Foundation, Museum of Fine Arts, Berne.
Sources and Acknowledgments
Project Physics Reader 4
1. Letter from Thomas Jefferson, June, 1799, from
Scripta Mathematica, Volume 1, 1932, pages
88-90. Reprinted with permission from the
Manuscript Collection of Teachers College
Library, Columbia University. Also in Science and
the Common Understanding by J. Robert
Oppenheimer. Reprinted with permission of
Simon and Schuster, Inc.
2. On the Method of Theoretical Physics by Albert
Einstein from Essays in Science, pages 12-21,
Philosophical Library, New York, copyright 1934
by Estate of Albert Einstein. Reprinted with
permission.
3. Systems, Feedback and Cybernetics from
Introduction to Natural Science by V. Lawrence
Parsegian, Alan S. Meltzer, Abraham S. Luchins
and K. Scott Kinerson, copyright © 1968 by
Academic Press, Inc. Reprinted with permission.
4. Velocity of Light from Studies In Optics by A. A.
Michelson, copyright 1927 by The University of
Chicago Press. Reprinted with permission.
5. Popular Application of Polarized Light from
Polarized Light (Momentum #7) by William A.
Shurcliff and Stanley S. Ballard, copyright © 1964
by Litton Educational Publishing Inc. Reprinted
with permission of Van Nostrand Reinhold
Company.
6. Eye and Camera by George Wald from Scientific
American, August 1950, copyright © 1950 by
Scientific American, Inc. Reprinted with permis-
sion. All rights reserved.
7. The Laser — What It Is and Does from The Story
of the Laser by John M. Carroll, copyright © 1964
by E. P. Dutton & Co., Inc., Reprinted with permis-
sion of E. P. Dutton & Co., Inc., and Souvenir
Press Ltd.
8. A Simple Electric Circuit: Ohm's Law from The
New College Physics: A Spiral Approach by
Albert V. Baez. W. H. Freeman and Company,
copyright© 1967. Reprinted with permission.
9. The Electronic Revolution from Voices from the
Sky by Arthur C. Clarke, Harper & Row, Publishers,
New York, copyright © 1965. "The Electronic
Revolution," copyright © 1962 by The New York
Times Company. Reprinted with permission of the
author and his agent, Scott Meredith Literary
Agency, and David Higham Associates. Ltd .
10. The Invention of the Electric Light by Matthew
Josephson from Scientific American, November
1959, copyright © 1959 by Scientific American,
Inc. Reprinted with permission. All rights reserved.
11. High Fidelity from Reproduction of Sound by
Edgar Villchur, copyright © 1962 by Acoustic
Research, Inc., Cambridge, Massachusetts.
Copyright © 1965 by Dover Publications, Inc.
Reprinted with permission.
12. The Future of Direct Current Power Transmission
by N. L. Allen from Endeavor, Volume CCVI, No.
97, Imperial Chemical Industries Limited, London.
Reprinted with permission.
13. James Clerk Maxwell by James R. Newman,
Part II, from Scientific American, June 1955,
copyright © 1955 by Scientific American, Inc.
Reprinted with permission. All rights reserved.
14. On the Induction of Electric Currents from
A Treatise on Electricity and Magnetism by James
Clerk Maxwell, Volume 2, 1873, The Clarendon
Press, Oxford. Reprinted with permission.
15. The Relationship of Electricity and Magnetism
from Faraday, Maxwell, and Kelvin, by D. K. C.
MacDonald, copyright © 1964 by Educational
Services Incorporated. Reprinted with permission
of Doubleday & Company, Inc. (Science Study
Series.)
16. The Electromagnetic Field from The Evolution of
Physics by Albert Einstein and Leopold Infeld.
Published by Simon and Schuster, copyright©
1961. Reprinted with permission.
17. Radiation Belts Around the Earth by James A.
Van Allen from Scientific American, March 1959.
Copyright © 1959 by Scientific American, Inc.
Reprinted with permission. All rights reserved.
Available separately at 200 each as Offprint No.
248 from W. H. Freeman and Company, 666
Market Street, San Francisco, California 94104.
18. A Mirror for the Brain from The Living Brain by
W. Grey Walter, copyright 1953, © 1963 by W. W.
Norton & Company, Inc. Reprinted with permission
of W. W. Norton & Company, Inc., and Gerald
Duckworth & Co. Ltd.
19. Scientific Imagination from The Feynman Lectures
on Physics by Richard P. Feynman, Robert B.
Leighton and Matthew L. Sands, copyright©
1964 by Addison-Wesley Publishing Company,
Inc. Reprinted with permission.
20. Lenses and Optical Instruments from PSSC
Physics, D. C. Heath and Company, Boston.
Copyright © 1965 by Educational Services
Incorporated. Reprinted with permission.
21. Baffled, by Keith Waterhouse from Punch, July 10,
1968, copyright © 1968 Punch, London.
in
This is not a physics textbook. Rather, it is a physics
reader, a collection of some of the best articles and
book passages on physics. A few are on historic events
in science, others contain some particularly memorable
description of what physicists do; still others deal with
philosophy of science, or with the impact of scientific
thought on the imagination of the artist.
There are old and new classics, and also some little-
known publications; many have been suggested for in-
clusion because some teacher or physicist remembered
an article with particular fondness. The majority of
articles is not drawn from scientific papers of historic
importance themselves, because material from many of
these is readily available, either as quotations in the
Project Physics text or in special collections.
This collection is meant for your browsing. If you follow
your own reading interests, chances are good that you
will find here many pages that convey the joy these
authors have in their work and the excitement of their
ideas. If you want to follow up on interesting excerpts,
the source list at the end of the reader will guide you
for further reading.
Reader 4
Table of Contents
1 Letter from Thomas Jefferson 1
June 1799
2 On the Method of Theoretical Physics 5
Albert Einstein
3 Systems, Feedback, Cybernetics 1 5
V. Lawrence Parsegian, Alan S. Meltzer, Abraham S. Luchins, K. Scott Kinerson
4 Velocity of Light 51
A. A. Michelson
5 Popular Applications of Polarized Light 69
William A. Shurcliff and Stanley Ballard
6 Eye and Camera 89
George Wald
7 The Laser— What It Is and Does 99
J. M. Carroll
8 A Simple Electric Circuit: Ohm's Law 143
Albert V. Baez
9 The Electronic Revolution 155
Arthur C. Clarke
1 0 The Invention of the Electric Light 162
Matthew Josephson
11 High Fidelity 175
Edgar Villchur
1 2 The Future of Direct Current Power Transmission 191
N. L. Allen
13 James Clerk Maxwell, Part II 195
James R. Newman
VI
14 On the Induction of Electric Currents
James Clerk Maxwell
229
1 5 The Relationship of Electricity and Magnetism
D. K. C. MacDonald
233
1 6 The Electromagnetic Field
Albert Einstein and Leopold Infeld
241
17 Radiation Belts Around the Earth
James Van Allen
249
18 A Mirror for the Brain
W. Grey Walter
259
19 Scientific Imagination
Richard P. Feynman, Robert B. Leighton, and Matthew Sands
285
20 Lenses and Optical Instruments
Physical Science Study Committee
289
21 Baffled!
Keith Waterhouse
301
VII
A great American writes about the significant role
of science in the education of the individual and in
the creation of American society.
Letter from Thomas Jefferson
June 1799
Monticello June 18. 99.
DEAR SIR,
I have to acknolege the reciept of your favor of
May 14. in which you mention that you have finished
the 6. first books of Euclid, plane trigonometry, sur-
veying and algebra and ask whether I think a further
pursuit of that branch of science would be useful to
you. There are some propositions in the latter books
of Euclid, and some of Archimedes, which are useful,
and I have no doubt you have been made acquainted
with them. Trigonometry, so far as this, is most
valuable to every man, there is scarcely a day in
which he will not resort to it for some of the purposes
of common life; the science of calculation also is
indispensible as far as the extraction of the square and
cube roots ; Algebra as far as the quadratic equation
and the use of logarithms are often of value in ordi-
nary cases ; but all beyond these is but a luxury ; a
delicious luxury indeed ; but not to be indulged in by
one who is to have a profession to follow for his sub-
sistence. In this light I view the conic sections,
curves of the higher orders, perhaps even spherical
trigonometry, Algebraical operations beyond the
2d dimension, and fluxions. There are other
branches of science however worth the attention of
every man : Astronomy, botany, chemistry, natural
philosophy, natural history, anatomy. Not indeed
to be a proficient in them ; but to possess their general
principles and outlines, so as that we may be able to
amuse and inform ourselves further in any of them as
we proceed through life and have occasion for them.
Some knowlege of them is necessary for our character
as well as comfort. The general elements of astro-
nomy and of natural philosophy are best acquired at
an academy where we can have the benefit of the
instruments and apparatus usually provided there:
but the others may well be acquired from books
alone as far as our purposes require. I have indulged
myself in these observations to you, because the
evidence cannot be unuseful to you of a person who
has often had occasion to consider which of his
acquisitions in science have been really useful to him
in life, and which of them have been merely a matter
of luxury.
I am among those who think well of the human
character generally. I consider man as formed for
society, and endowed by nature with those disposi-
tions which fit him for society. I believe also, with
Condorcet, as mentioned in your letter, that his mind
is perfectible to a degree of which we cannot as yet
form any conception. It is impossible for a man who
takes a survey of what is already known , not to see
what an immensity in every branch of science yet
remains to be discovered, and that too of articles to
which our faculties seem adequate. In geometry and
calculation we know a great deal. Yet there are
some desiderata. In anatomy great progress has
been made; but much is still to be acquired. In
natural history we possess knowlege; but we want
a great deal. In chemistry we are not yet sure of
the first elements. Our natural philosophy is in a
very infantine state; perhaps for great advances in
it, a further progress in chemistry is necessary.
Surgery is well advanced ; but prodigiously short of
what may be. The state of medecine is worse than
that of total ignorance. Could we divest ourselves of
Letter from Thomas Jefferson
every thing we suppose we know in it, we should start
from a higher ground and with fairer prospects.
From Hippocrates to Brown we have had nothing
but a succession of hypothetical systems each having
it's day of vogue, like the fashions and fancies of
caps and gowns, and yielding in turn to the next
caprice. Yet the human frame, which is to be the
subject of suffering and torture under these learned
modes, does not change. We have a few medecines,
as the bark, opium, mercury, which in a few well
defined diseases are of unquestionable virtue: but
the residuary list of the materia medica, long as it is,
contains but the charlataneries of the art ; and of the
diseases of doubtful form, physicians have ever had a
false knowlege, worse than ignorance. Yet surely
the list of unequivocal diseases and remedies is
capable of enlargement ; and it is still more certain
that in the other branches of science, great fields are
yet to be explored to which our faculties are equal,
and that to an extent of which we cannot fix the
limits. I join you therefore in branding as cowardly
the idea that the human mind is incapable of further
advances. This is precisely the doctrine which the
present despots of the earth are inculcating, and
their friends here re-echoing ; and applying especially
to religion and politics ; ' that it is not probable that
any thing better will be discovered than what was
known to our fathers '. We are to look backwards
then and not forwards for the improvement of
science, and to find it amidst feudal barbarisms and
the fires of Spital-fields. But thank heaven the
American mind is already too much opened, to listen
to these impostures ; and while the art of printing is
left to use, science can never be retrograde; what is
once acquired of real knowlege can never be lost.
To preserve the freedom of the human mind then and
freedom of the press, every spirit should be ready to
devote itself to martyrdom ; for as long as we may
think as we will, and speak as we think, the condition
of man will proceed in improvement. The genera-
tion which is going off the stage has deserved well of
mankind for the struggles it has made, and for having
arrested that course of despotism which had over-
whelmed the world for thousands and thousands of
years. If there seems to be danger that the ground
they have gained will be lost again, that danger comes
from the generation your contemporary. But that
the enthusiasm which characterises youth should lift
it's parracide hands against freedom and science
would be such a monstrous phaenomenon as I can-
not place among possible things in this age and this
country. Your college at least has shewn itself
incapable of it; and if the youth of any other place
have seemed to rally under other banners it has been
from delusions which they will soon dissipate. I
shall be happy to hear from you from time to time,
and of your progress in study, and to be useful to
you in whatever is in my power; being with sincere
esteem Dear Sir
Your friend & servt
Th : Jefferson
Einstein discusses some of the factors that lead to a
scientific theory.
On the Method of Theoretical Physics
Albert Einstein
An essay— 1934.
If you want to find out anything from the theoretical
physicists about the methods they use, I advise you
to stick closely to one principle : don't listen to their
words, fix your attention on their deeds. To him
who is a discoverer in this field the products of his
imagination appear so necessary and natural that he
regards them, and would like to have them regarded
by others, not as creations of thought but as given
realities.
These words sound like an invitation to you to
walk out of this lecture. You will say to yourselves,
the fellow's a working physicist himself and ought
therefore to leave all questions of the structure of
theoretical science to the epistemologists.
Against such criticism I can defend myself from
the personal point of view by assuring you that it
is not at my own instance but at the kind invitation
of others that I have mounted this rostrum, which
serves to commemorate a man who fought hard all
his life for the unity of knowledge. Objectively, how-
ever, my enterprise can be justified on the ground
that it may, after all, be of interest to know how one
who has spent a life-time in striving with all his
might to clear up and rectify its fundamentals looks
upon his own branch of science. The way in which he
regards its past and present may depend too much
on what he hopes for the future and aims at in the
present; but that is the inevitable fate of anybody
who has occupied himself intensively with a world
of ideas. The same thing happens to him as to the
historian, who in the same way, even though perhaps
unconsciously, groups actual events around ideals
which he has formed for himself on the subject of
human society.
Let us now cast an eye over the development of
the theoretical system, paying special attention to
the relations between the content of the theory
and the totality of empirical fact. We are concerned
with the eternal antithesis between the two insep-
arable components of our knowledge, the empirical
and the rational, in our department.
We reverence ancient Greece as the cradle of
western science. Here for the first time the world
witnessed the miracle of a logical system which pro-
ceeded from step to step with such precision that
every single one of its propositions was absolutely
indubitable — I refer to Euclid's geometry. This ad-
mirable triumph of reasoning gave the human intel-
lect the necessary confidence in itself for its subsequent
achievements. If Euclid failed to kindle your youth-
ful enthusiasm, then you were not born to be a
scientific thinker.
But before mankind could be ripe for a science
which takes in the whole of reality, a second funda-
On the Method of Theoretical Physics
mental truth was needed, which only became common
property among philosophers with the advent of Kep-
ler and Galileo. Pure logical thinking cannot yield
us any knowledge of the empirical world ; all knowl-
edge of reality starts from experience and ends in
it. Propositions arrived at by purely logical means
are completely empty as regards reality. Because
Galileo saw this, and particularly because he drummed
it into the scientific world, he is the father of modern
physics — indeed, of modern science altogether.
If, then, experience is the alpha and the omega of
all our knowledge of reality, what is the function of
pure reason in science?
A complete system of theoretical physics is made
up of concepts, fundamental laws which are supposed
to be valid for those concepts and conclusions to be
reached by logical deduction. It is these conclusions
which must correspond with our separate experiences ;
in any theoretical treatise their logical deduction
occupies almost the whole book.
This is exactly what happens in Euclid's geometry,
except that there the fundamental laws are called
axioms and there is no question of the conclusions
having to correspond to any sort of experience. If,
however, one regard Euclidean geometry as the sci-
ence of the possible mutual relations of practically
rigid bodies in space, that is to say, treats it as a
physical science, without abstracting from its original
empirical content, the logical homogeneity of geometry
and theoretical physics becomes complete.
We have thus assigned to pure reason and ex-
perience their places in a theoretical system of physics.
The structure of the system is the work of reason ; the
empirical contents and their mutual relations must
find their representation in the conclusions of the
theory. In the possibility of such a representation lie
the sole value and justification of the whole system,
and especially of the concepts and fundamental prin-
ciples which underlie it. These latter, by the way, are
free inventions of the human intellect, which cannot
be justified either by the nature of that intellect or
in any other fashion a priori.
These fundamental concepts and postulates, which
cannot be further reduced logically, form the essential
part of a theory, which reason cannot touch. It is the
grand object of all theory to make these irreducible
elements as simple and as few in number as possible,
without having to renounce the adequate representa-
tion of any empirical content whatever.
The view I have just outlined of the purely fictitious
character of the fundamentals of scientific theory
was by no means the prevailing one in the eighteenth
or even the nineteenth century. But it is steadily
gaining ground from the fact that the distance in
thought between the fundamental concepts and laws
on one side and, on the other, the conclusions which
have to be brought into relation with our experience
grows larger and larger, the simpler the logical struc-
ture becomes — that is to say, the smaller the number
of logically independent conceptual elements which
are found necessary to support the structure.
Newton, the first creator of a comprehensive,
On the Method of Theoretical Physics
workable system of theoretical physics, still believed
that the basic concepts and laws of his system could
be derived from experience. This is no doubt the
meaning of his saying, hypotheses non fingo.
Actually the concepts of time and space appeared
at that time to present no difficulties. The concepts
of mass, inertia and force, and the laws connecting
them seemed to be drawn directly from experience.
Once this basis is accepted, the expression for the
force of gravitation appears derivable from experi-
ence, and it was reasonable to hope for the same in
regard to other forces.
We can indeed see from Newton's formulation of
it that the concept of absolute space, which comprised
that of absolute rest, made him feel uncomfortable ;
he realized that there seemed to be nothing in ex-
perience corresponding to this last concept. He was
also not quite comfortable about the introduction of
forces operating at a distance. But the tremendous
practical success of his doctrines may well have pre-
vented him and the physicists of the eighteenth and
nineteenth centuries from recognizing the fictitious
character of the foundations of his system.
The natural philosophers of those days were, on
the contrary, most of them possessed with the idea
that the fundamental concepts and postulates of
physics were not in the logical sense free inventions
of the human mind but could be deduced from ex-
perience by "abstraction" — that is to say by logical
means. A clear recognition of the erroneousness of
this notion really only came with the general theory
of relativity, which showed that one could take ac-
count of a wider range of empirical facts, and that
too in a more satisfactory and complete manner, on
a foundation quite different from the Newtonian.
But quite apart from the question of the superiority
of one or the other, the fictitious character of funda-
mental principles is perfectly evident from the fact
that we can point to two essentially different prin-
ciples, both of which correspond with experience to
a large extent ; this proves at the same time that
every attempt at a logical deduction of the basic con-
cepts and postulates of mechanics from elementary
experiences is doomed to failure.
If, then, it is true that this axiomatic basis of theo-
retical physics cannot be extracted from experience
but must be freely invented, can we ever hope to
find the right way? Nay more, has this right way any
existence outside our illusions? Can we hope to be
guided in the right way by experience when there
exist theories (such as classical mechanics) which to
a large extent do justice to experience, without
getting to the root of the matter? I answer without
hesitation that there is, in my opinion, a right way,
and that we are capable of finding it. Our experience
hitherto justifies us in believing that nature is the
realization of the simplest conceivable mathematical
ideas. I am convinced that we can discover by means
of purely mathematical constructions the concepts
and the laws connecting them with each other, which
furnish the key to the understanding of natural phe-
nomena. Experience may suggest the appropriate
10
On the Method of Theoretical Physics
mathematical concepts, but they most certainly cannot
be deduced from it. Experience remains, of course,
the sole criterion of the physical utility of a mathe-
matical construction. But the creative principle resides
in mathematics. In a certain sense, therefore, I hold
it true that pure thought can grasp reality, as the
ancients dreamed.
In order to justify this confidence, I am compelled
to make use of a mathematical conception. The phys-
ical world is represented as a four-dimensional con-
tinuum. If I assume a Riemannian metric in it and
ask what are the simplest laws which such a metric
system can satisfy, I arrive at the relativist theory
of gravitation in empty space. If in that space I
assume a vector-field or an anti-symmetrical tensor-
field which can be inferred from it, and ask what
are the simplest laws which such a field can satisfy,
I arrive at Clerk Maxwell's equations for empty space.
At this point we still lack a theory for those parts
of space in which electrical density does not disappear.
De Broglie conjectured the existence of a wave field,
which served to explain certain quantum properties
of matter. Dirac found in the spinors field-magni-
tudes of a new sort, whose simplest equations enable
one to a large extent to deduce the properties of the
electron. Subsequently I discovered, in conjunction
with my colleague, that these spinors form a special
case of a new sort of field, mathematically connected
with the four-dimensional system, which we called
"semi vectors." The simplest equations to which such
semivectors can be reduced furnish a key to the
understanding of the existence of two sorts of ele-
mentary particles, of different ponderable mass and
equal but opposite electrical charge. These semivectors
are, after ordinary vectors, the simplest mathematical
fields that are possible in a metrical continuum of
four dimensions, and it looks as if they described, in
an easy manner, certain essential properties of elec-
trical particles.
The important point for us to observe is that all
these constructions and the laws connecting them can
be arrived at by the principle of looking for the mathe-
matically simplest concepts and the link between
them. In the limited nature of the mathematically
existent simple fields and the simple equations pos-
sible between them, lies the theorist's hope of grasp-
ing the real in all its depth.
Meanwhile the great stumbling-block for a field-
theory of this kind lies in the conception of the
atomic structure of matter and energy. For the theory
is fundamentally non-atomic in so far as it operates
exclusively with continuous functions of space, in
contrast to classical mechanics, whose most impor-
tant element, the material point, in itself does justice
to the atomic structure of matter.
The modern quantum theory in the form associated
with the names of de Broglie, Schrodinger, and
Dirac, which operates with continuous functions, has
overcome these difficulties by a bold piece of inter-
pretation which was first given a clear form by Max
Born. According to this, the spatial functions which
appear in the equations make no claim to be a mathe-
12
On the Method of Theoretical Physics
matical model of the atomic structure. Those func-
tions are only supposed to determine the mathematical
probabilities of the occurrence of such structures if
measurements were taken at a particular spot or in a
certain state of motion. This notion is logically un-
objectionable and has important successes to its
credit. Unfortunately, however, it compels one to use
a continuum the number of whose dimensions is not
that ascribed to space by physics hitherto (four) but
rises indefinitely with the number of the particles
constituting the system under consideration. I cannot
but confess that I attach only a transitory importance
to this interpretation. I still believe in the possibility
of a model of reality — that is to say, of a theory which
represents things themselves and not merely the
probability of their occurrence.
On the other hand it seems to me certain that we
must give up the idea of a complete localization of
the particles in a theoretical model. This seems to
me to be the permanent upshot of Heisenberg's
principle of uncertainty. But an atomic theory in the
true sense of the word (not merely on the basis of
an interpretation) without localization of particles
in a mathematical model, is perfectly thinkable. For
instance, to account for the atomic character of elec-
tricity, the field equations need only lead to the
following conclusions: A portion of space (three-
dimensional) at whose boundaries electrical density
disappears everywhere, always contains a total elec-
trical charge whose size is represented by a whole
number. In a continuum-theory atomic characteristics
13
would be satisfactorily expressed by integral laws
without localization of the formation entity which
constitutes the atomic structure.
Not until the atomic structure has been successfully
represented in such a manner would I consider the
quantum-riddle solved.
14
One process can cause another; that one in turn, can be
the cause of a further sequence of events— including the
modification of the original process itself. This article is a
primer to basic ideas in applied science, engineering, and
information theory.
3 Systems, Feedback, Cybernetics
V. Lawrence Parsegian, Alan S. Meltzer, Abraham S. Luchins,
K. Scott Kinerson
From the textbook, Introduction to Natural Science, 1 968.
the READER will recall that following
the quotation from Teilhard de Chardin
in Chapter 1, we proposed extending the
scope of our interests to include analysis
of relationship and interrelationship of
natural phenomena to each other. We
have come to a point that requires a
more formal development of such inter-
relationships.
6.1 Extension of "systems"
One of the accomplishments of the New-
tonian period was the strengthening
of the concept that in material or physical
situations at least, things do not happen
without a causing force. A stone does not
hegin to move or come to a stop of its
own volition. In this chapter we shall
utilize that concept, but with three ex-
tensions.
The first extension takes into account
the fact that in most situations surround-
ing an event (such as the hurling of a
stone), the immediate event is itself part
of a larger situation or system that in-
cludes various other articulating parts
or related events. (That is, there is a
person who throws the stone, and the
throwing has relation to some cause or
purpose.)
The second extension may perhaps
be thought of as related to the action-
reaction principle, namely, that within
the context of the system involving an
event (a stone is thrown) there is often a
feedback effect (for example, the one at
whom the stone is thrown may hurl it
back).
The third extension includes in the
system both material things (stones) and
human beings along with biological
processes and the less tangible thought
processes.
What do we mean by the term system?
We might refer to the weight suspended
15
from a spring as a system that executes
simple harmonic motion. The governor
that controls the speed of an engine is a
control system. We also speak of a sys-
tem of highways, the economic system
of a nation, a system of thought, and of
many others. The combination locks that
protect the vault of a savings bank make
up a protective system, but this can also
be said to be only a subsystem of the
banking institution. The banking institu-
tion is itself only a subsystem within the
larger community economics, and the
latter is a subsystem of national eco-
nomics. The chain of larger and larger
subsystems, or the nesting of subsystems
within larger subsystems, may lead to
very complex assemblies and relation-
ships.
While an accurate, all-encompassing
definition for the term is not easy to give,
we can note a few of the characteristics
that are usually present in what we call
a system:
(1) A system is likely to have two or
more parts, elements, or aspects, which
tend to have some functional relation to
each other (like the bolt and key of the
lock, or the president and staff of the
bank).
(2) Because systems are usually sub-
systems of larger units it is usually help-
ful (and often necessary) to confine one's
study to the smallest unit that encom-
passes the particular functional ele-
ments and interrelationships that are
under study. (For example, the locksmith
can quite properly repair a fault in the
lock system of the bank vaults without
considering the question of the merits
of socialism for the nation's banking
system.)
(3) A control system has within itself
regulatory functions for control of vari-
ables such as speed of a motor, the tem-
perature of a room, the price of commod-
ities, or international trade in narcotics.
(4) It is usually possible to identify an
"input" and an "output" portion (or
aspect) of a system. For example, a key
placed in a lock and turned (input) will
cause the bolts to move (output); or an
order from a president of an industrial
firm (input) can double the selling price
of its commercial products (output). We
shall find, however, that most systems
have more than one form of input, as well
as a variety of functional relationships
that produce quite varied output.
(5) Usually (nearly always in systems
that include regulatory functions) there
is some form of feedback from the output
to the input, which may greatly modify
the net output of the system. [For exam-
ple, when the selling prices of the com-
mercial products of paragraph (4) were
doubled, the consumers could have
initiated strong feedback by refusing to
buy the products; and the industry's
board of directors could have exerted
even stronger feedback by firing the
president and hiring another who would
hold the prices at a more acceptable
level.] The role of feedback will be given
considerable attention in the discussion
that follows.
We shall now turn to a more detailed
introduction to systems, feedback, and
control.
6.2 Cyclic character of
natural phenomena
In Chapter 5 we learned that a mass sus-
pended from a spring executes simple
harmonic motion when displaced slightly
from its equilibrum position. When the
motion was recorded on a moving sheet
of paper (to illustrate the motion as a
function of time), the oscillations were
16
Systems, Feedback, Cybernetics
recorded as sine or cosine waveforms.
It was shown that the motion was initiated
when potential energy was added to the
system of weight and spring (by manually
raising the weight from its rest position,
against the pull of gravity, or by pulling
it down and extending the spring). In
either case, the pull of gravity or the pull
of the spring alternately introduced a
restoring force, which tended to return
the displaced mass to its original position
(Fig. 5.24). But since force applied to
mass accelerates the mass and thereby
increases its velocity (Eq. 5-1), by the
time the mass reached the "zero" or initial
position it had acquired so much velocity
(because the potential energy we added
manually had become kinetic energy at
that point) that the mass moved past the
zero point to the other extreme. There
would have been few or no oscillations at
all, on the other hand, if the weight had
been subjected to so much frictional drag
that the added (potential) energy was
lost as heat.f (This might have been the
case if the weight moved in a viscous
liquid.)
What about cyclic behavior in other
phenomena of nature? A very common
form can be demonstrated in electric
circuits in which the electric energy
rapidly passes back and forth between
parts of an oscillating circuit until the
electric energy dissipates as heat or
radiates away from the circuit (as in the
transmission of radio waves).
We shall find that there can be many
forms of oscillatory behavior when a
t We shall learn in Chapter 10 that the kinetic
energy of the system goes into faster, random
motion of the molecules that make up the parts
of the system. The increased molecular motion
raises the temperature of the parts of the sys-
tem, as though it were heated by a flame.
There is therefore a correspondence or equiv-
alence between the energy in a flame and
mechanical motion of the system.
"disturbance" changes the energy level
of a system and introduces a restoring
force that causes the energy to convert
to another form rather than completely
dissipate into the heat energy of the en-
vironment. The term energy may apply
not only to mechanical, electrical, or
chemical characters in physical systems,
but also to institutional and personal
pressures in social situations.
Let us now go to other phenomena that
show cyclic or periodic variation. (See
Figs. 6.1(a) through 6.1(d), for graphical
examples of such cyclic variations.) We
might utilize various sensing devices to
record changes in the temperature of an
air-conditioned room as a function of
time, the height of the tides of the sea,
wind velocity, the automobile traffic on
a road, rainfall, the movements of a tall
building or of the long span of a bridge,
or the temperature of the earth. We might
also look up past statistics on wheat
production, the stock market, attendance
at church, tourist travel, populations of
animals, or the length of women's skirts,
and plot these in graph form as function
of time. We would find that many phe-
nomena in nature and in animal or social
activity have variations of an oscillating
character (Fig. 6.1). It can be demon-
strated that in all these situations which
show oscillations about some average
point, there is present a restoring force
that comes into play whenever there
is energy change in a system. To be sure,
the magnitude and shapes of these oscil-
lations and waves vary considerably from
the sine waves we observe with a weight
on a spring. The periods may vary from
10-15 sec in the case of light waves, to
several hours for the period of the tides,
and to many years in the case of other
cycles of nature and of some social cus-
toms. Nevertheless, all are subject to
some common influence, not the least of
17
3rd interglacial
period
1st glacial 2nd glacial
period period
3rd glacial 4th glacial
period period
6 5 4 3 2 10
Time in hundreds of thousands of years ago
Fig. 6.1. (a) Cyclic temperature
variations during the ice ages.
Current theory attributes these
long, slow temperature variations
to relatively minor changes in the
atmospheric carbon dioxide
content (see Chapter 15, Sec. 2).
(Adapted from graph in G. H.
Drury, The Face of the Earth,
Penguin (Pelican book), pg. 157.)
140^
Fig. 6.1. (b) Cyclic varia-
tions in numbers of species
of Lepidoptera (butterflies
and moths) captured in light
traps at Woking, Surrey in
1948-49. The number of
different species of captured
reveals seasonal cyclic
variations that are obviously
related to weather condi-
tions. Note peaks in
successive Julys, when
Lepidoptera conditions are
ideal, and low values in
winter when conditions are
poor. (From C. B. Williams,
Patterns in the Balance of
Nature, Academic Press,
1964, pg. 159.)
-
2 120 =
0) —
Q.
o —
-
-
-
Q. —
0)
o 80-
i/> —
0) —
'o —
® fin
-
<n —
o _
S3 40^
JQ —
E
z 20-
-
i
i c
! <
N
- 1
3 -
» —
! 5
«« 1
f 1
C (j
h
3 2
5 j
! C
u
Z J
0
-1 L
3 i
L <
1
> <
h
0 !
> .:
! ■
5 "
-> -
>- t
-) <
t 0
0 c
3
>
->
V V
1948 1949
18
Systems, Feedback, Cybernetics
-6 »>.
8 ?
■n c
O 3
32
30
28
26
24
22
20
A
A
<\
r
yv
A
Y
^
A
/
\
A
V/
\-
-4
o
lO
01
Fig. 6.1. (c) Cyclic char-
acter of employment levels
in U.S. goods producing
industries, 1920-1960. Note
large amplitude cycles
superimposed on more
normal fluctuations as a
result of the depression in
the early 1930's and of
World War II during the
early 1940' s.
RflO
MM
1 1 1 1
MM
MM
MM
II 1 1
1 1
Fig. 6.1. (d) Cyclic
variation in value of
new construction of
religious buildings.
450 E
400 E
in ~
« 350 _
■3 300 E
g 250 E
o —
= 200-
2 =
150-
100 =
50 E
=
19
II II
20 19
MM
25 19
II II
30 19
MM
35 19
1 1 1 1
40 19
1 1 II
45 19
1 1
50
which is the fact that nature is dynamic
and in a state of continuous change, and
indeed that static situations represent
special and almost trivial aspects of nature
and of man.
Is the presence of some restoring force
sufficient assurance that a system will
experience only moderate oscillations
without going to extremes? Indeed it is
not, as we can learn from the dramatic
example of the failure of the Tacoma
Narrows suspension bridge of Tacoma,
Washington. When the bridge was opened
to traffic on July 1, 1940, there were
observed, in addition to the ordinary oscil-
lations of the bridge, some unexpected
19
Hand holding
spring
nitial amplitude
ot mass
Motion of hand
— Opposite motion
of mass
Motion of the mass
becomes reduced in
amplitude
Motion of hand
— "In phase" with
motion of mass
Motion of the mass
builds to destructive
proportions
Fig. 6.2. What happens when the mass of a spring is given some additional energy
by movement of the hand in two different phase relationships? In the center figure
the hand is moved upward when the mass is moving downward. At the right, the hand
is moved upward when the mass is also moving upward, causing the mass to take
large swings.
transverse (vertical) modes of vibration.
On November 7 a wind velocity of 40 to
45 mph made the vibrations so severe
that the bridge was closed to traffic, and
by 11:00 A.M. the main span collapsed, f
t A 4-minute film produced by the Ohio
State University and distributed by The Ealing
Corporation of Cambridge, Massachusetts,
gives the very dramatic story of the final oscil-
lation of the bridge prior to its collapse. Everv
reader should see this film and the variation
it offers of "simple harmonic motion" involving
the twisting and turning of this huge span of
steel and concrete. The new bridge that was
built on the original anchorages and tower
foundations included deep stiffening trusses
instead of girders, and has been entirely
successful.
6.3 How oscillations increase
despite restoring forces
It is not necessary to resort to the com-
plex behavior of the original Tacoma
Narrows bridge to see how a system may
have within it strong restoring forces
while yet experiencing oscillations that
increase in amplitude to the point of
destruction. The reader can duplicate
the phenomenon with the simple weight
on a spring as follows (Fig. 6.2):
Choose a weight and spring combina-
tion that gives an oscillatory period be-
tween i and i sec. Hold the spring firmly
and steady in your hand, and observe that
the weight executes the usual simple
20
Systems, Feedback, Cybernetics
harmonic motion, eventually coming to
a stop. Now prepare to move your hold-
ing hand up or down in synchronism
with the motion of the weight and with
two alternative movements.
First, raise your hand (about a half-inch
will do) whenever the weight is moving
downward, and lower it an equal distance
whenever the weight is moving up. With
a little analysis you can see that the
weight tends to reduce amplitude be-
cause the movements of your hand in-
crease the restoring force on the weight.
Note that the movement of your hand is
180 deg out of phase with the motion of
the weight.
Next, repeat the experiment with the
same up-and-down motion of your hand,
but now change the timing to be in phase
with the motion of the weight. That is,
move your hand upward when the weight
is moving upward, and downward when
the weight is moving downward. There
still is restoring force, and the weight
continues to oscillate up and down; but
now the amplitude of oscillations in-
creases until it becomes dangerous to
continue the experiment.
Why did the same amount of motion of
your hand have such opposite effects,
depending only on its phase relationship
to the motion of the weight? The reason
is that in the second case the increments
of energy that were introduced by each
in-phase motion of your hand tended to
add to and increase the energy of the
system represented by the spring and
weight. Conversely, the hand motion that
was completely out of phase with the
motion of the weight detracted from the
energy of the system, t
t The reader is urged to perform this experi-
ment and to attempt a careful analysis of the
various factors (energies and forces) that be-
come involved in the two cases. For example,
We can now extend this experiment to
apply to the early Tacoma Narrows bridge
experience. Obviously, the energy of the
wind became converted to energy of
oscillation of the bridge. Why did the
wind energy not become absorbed in
the concrete and steel of the bridge? Un-
doubtedly much of it did become ab-
sorbed and changed to heat energy, but
not all of it. Apparently when the wind
blew to produce a movement of the span
at some point along the bridge, the con-
ditions were just right to cause this move-
ment to act as a traveling wave, which on
backward reflection returned to the same
point in just the right phase to support
(rather than oppose) a new movement at
that point, caused by the continued blow-
ing of the wind. Had the physical struc-
ture of the bridge been different in length
or mass, the returning wave could have
opposed (out of phase with) any new
movement at A, and thus would have
added to the stability of the system.
We see, therefore, that for a system to
be stable, the relationship of the forces
and time characteristics must be such that
the amplitude and energy of the system
will not increase. This calls for special
attention with respect to the phase re-
lationships that obtain between feed-
back of energy from one part of the sys-
tem to another part. When the feedback
opposes the direction of the initial change
that produced the feedback, the system
tends to be stable. In contrast, when the
returning feedback of energy supports
the direction of initial change, the system
tends to add to the initial energy gain and
to be unstable. This means we must
delve into the theory of system control.
in the second case the increments of energy are
added to the spring-weight system. Where
does the hand energy go in the first case?
21
6.4 Modifying cyclic changes:
controls
While most fluctuations of nature go their
own way without inviting human concern,
there are some important cases in which
it becomes necessary to interfere, that is,
to modify the natural pattern or to control
or hold the fluctuations to smaller
changes. For example, the farmer may
not want to depend entirely on natural
rainfall to assure a good crop, so he in-
tervenes by irrigating the fields when
there is not enough rainfall. Because in
the course of the year there are wide
fluctuations in the temperature of the
earth, he installs a control system in his
home to keep the temperature within
comfortable limits.
Many types of controls are involved
in our daily life. We shall learn that the
human body has a remarkable control
system to maintain its own temperature
within very close limits. The body's
motor functions, by which we move our
arms and legs in an accurate and deter-
mined manner, are possible only because
of the operation of fine control systems.
Industrial production relies heavily on
control of temperature, pressure, chemi-
cal composition, and similar factors. The
application of control principles extends
to community and national life. Despite
their variety, we shall find that there are
some common characteristics among
them. Also, within a specific control
system there can be intermixed a wide
variety of elements of widely different
types. Take, for example, the very com-
mon experience of driving an automobile.
Here, the steering control allows the
driver to follow the curvature of the road
effectively, and many other electrome-
chanical parts as well in the motor and
transmission systems affect the driving
operation. But we shall learn before long
that nearly every aspect of the driver's
being — his metabolism, muscle and nerve
action, his thinking process — and the life
of his community are all parts of the sys-
tem that encompasses the simple driving
experience.
6.5 Introduction to on-off control
We return to the harmonic motion of the
weight suspended from a spring and note
that, so far, we have neither tried to re-
strict the amplitude of the motion nor put
the movement to some useful application.
In each assembly the added energy is
converted and reconverted from kinetic
energy to potential energy and then back
again to kinetic energy. (If there were no
frictional losses, the motion would con-
tinue forever, since the system would
then be self-contained, that is, a closed
or isolated system that neither receives
energy from nor gives energy to the out-
side.) Such systems have limited value
except as one may use them in a clock or
metronome to tell time from the os-
cillations.!
If there were no frictional or other loss
of energy from the system, the motion
would have a periodicity of T seconds.
Since friction is present, the oscillations
become continually smaller in magni-
tude, and the period of each cycle be-
comes slightly longer (T + AT) until the
mechanical energy dissipates as heat
energy and the movement ceases al-
together (Fig. 6.3). In general, friction or
damping is likely to make a system more
stable.
We can design an oscillator to do some-
f Of course, as any such device requires
periodic additions ot energy to the driving
springs, and therefore the person who winds
the spring becomes part of the system.
22
Systems, Feedback, Cybernetics
3 1
"5.
E '
i
i
1
""^^v
,
\ Time—*- /
'
1
I —
w. 1
T + AT sec-
Fig. 6.3. How the period of
simple harmonic motion
changes when there is
friction in the system. (The
period of seconds increases
to T + AT sec, while the
amplitude of motion
decreases.)
thing more by adding an electric switch
so that the dropping weight sends an
electrical signal to some device. As we
know from common experience, the
simple operation of an electric switch
can initiate (or trigger) many motor or
relay functions that bring into play the
vast energy resources of electric power-
generating stations. Figure 6.4 illustrates
the relationship between input and out-
put, with a transform function that relates
the two along with a source of energy.
Suppose that we incorporate such an
electric switch as part of a control system
for automatically filling a bucket with
water. Figure 6.5 illustrates how the
dropping pail signals that the pail is full
and also turns off the stream of water.
This becomes a simple on-off control
system in which the electrical signal
provides a feedback function as part
of the control system. (Later we shall
introduce the idea that the feedback also
represents information.)
We examine this process of filling the
bucket in a little more detail. When the
water flows into the bucket at a very slow
rate, the bucket settles slowly and the
signal switch has time to stop the flow
of water and bring the bucket to a gentle
stop. This is shown as curve A of Fig. 6.6,
which shows very little dropping of the
bucket below the desired level (that is,
there is very little overshoot beyond the
desired control point). The behavior
becomes quite different when the water
flows into the bucket at a rapid rate,
however. The switch operates as it did
before, but the rapid dropping of the
bucket develops enough momentum to
overshoot the desired final position by a
substantial amount. The bucket will
oscillate violently above and below the
desired control height for some time and
the switch will open and close erratically
(curve B). In fact, if the response rates
and delays in the switching and valve
devices should turn out to be particularly
23
Energy source to make
amplification possible
lllll
A/ = input to system
This box establishes
relationship between output
and input — the transfer
function /(/)
bO = output of system
>_
(For example, position of (For example, automobile motor
accelerator pedal of an develops large power to correspond
automobile or position of to accelerator position, or electric
an electrical switch) circuit brings large electric energy
into play)
Fig. 6.4. How a small input change (such as the operation of an electric switch) can
bring into play sources of energy and thereby produce an output that may be quite
different inform and magnitude from the input. Each such conversion can be referred
to as involving a transformation (transfer function or transform function).
unsuitable, the water would be turned
on and off in such erratic manner as to
recall the sad fate of the Tacoma Narrows
bridge; see curve B, dotted line, Fig. 6.6.
In the case of room-temperature control,
the thermostat is likely to be kept at one
temperature (for comfort), say, around
72°F. In the case of the baking oven, the
temperature setting will vary with the
requirements for baking a cake or roast-
ing meat. In either case, the temperature
will vary (or hunt) around the set control
point. The hunting or oscillations can be
decreased if the rate of heat input is slow.
But this would increase the time needed
to bring the room or oven to the desired
temperature. With on-off control, the
heating unit becomes fully hot whenever
the control switch turns it on. By the time
the temperature at the thermostat reaches
the desired temperature to turn off the
heat, the region of the heater units be-
comes much hotter than necessary, and
this excess heat drives the temperature
well above the desired temperature. A
similar delay in reactivating the heating
unit as the temperature drops below the
desired level causes continual hunting
above and below the desired temperature.
We shall appreciate more and more, as
we examine more cases, that the "control"
of a variable rarely results in an exact
holding of the variable to the desired
control value. Nearly always, the variable
will hunt or vary about that control
value. Therefore, the function of a suc-
cessful control system is to hold the
variable within acceptable departures
from the desired control value.
6.6 Negative versus
positive feedback
In all the examples given above, while it
is clear that control at a point usually ends
up as hunting around that point, even this
24
Systems, Feedback, Cybernetics
'/////////////////////
Fig. 6.5. A simple system for controlling the
filling of a bucket.
Spring
Valve to control water flow by
electric motor control
To motor valve
Electric switch, designed to turn off
water valve when bucket drops
down to close switch
Fig. 6.6. How the bucket of Fig. 6.5 behaves:
Bucket A is filled slowly and settles gradually to
its final level after switch cuts off water flow.
Bucket B (solid line) is filled rapidly and over-
shoots final position, rebounds, and hunts for an
equilibrium position that is lower than that for
bucket A because extra water was added after the
first rebound above the switch-off level. With a
different spring tension for bucket B (dotted line),
the hunting may cause addition of sufficient
extra water on each cycle so that the amplitude
steadily increases until the system collapses.
Excessive hunting results in collision
with upper support and final collapse
of controlled aspect of system
bucket
25
degree of control is achieved only when
negative feedback is present. Thus, in
the case of the full bucket, the switch
turns off the water (since it was the
"water-on" condition that filled the
bucket). In the case of room-temperature
control (which we shall discuss presently
in detail), the heaters must be turned on
when the room temperature is too low,
and off when the temperature is too high.
The examples of feedback, as well as
the limitations of on-off (sometimes called
bang-bang) control can be illustrated
further by the example of a blind person
walking down a street with his cane. As
he progresses along the sidewalk the
tapping of his cane tells him when he is
too close to the buildings on the right.
This information, when processed
through his brain and muscle system,
serves as feedback to change his direction.
Since his movements have taken him too
far to the right, now he must move to the
left and therefore the feedback must be
negative. If the influence of feedback
were positive, it would support or add
to the original direction that took him t&
the right and would take him even farther
to the right and directly into the wall. He
now continues to the left until his cane
warns that he is too close to the curb at
the left. This information again converts
to become negative feedback, which will
oppose the move that carried him too far
to the left and thereby will restore his
direction until a new signal calls for new
action.
Our blind person can negotiate the
walk fairly well as long as his movements
are slow enough to give him time to
receive the signal from his tapping, to
interpret these, and to translate them into
suitable feedback influence. But now
suppose he tries to run down the same
sidewalk. Very soon his rate of receiving
and responding to signals would be in-
adequate, and he would be running in a
zigzag or colliding with obstacles.
Such an experience, which the reader
can himself check rather dramatically,
illustrates several features of control
that apply fairly generally, namely:
(1) Stable control requires the pres-
ence of negative feedback influences.
(2) Stable control of a variable to a
"fixed" point usually means maintain-
ing the variable so that is does not hunt
around the point beyond acceptable
limits.
(3) To be effective for the control of
any variable, the control system must
be designed to have response rates that
are suited for the specific application.
These and other characteristics of
control systems will be illustrated in the
following sections.
6.7 Driving an automobile
To illustrate further the limitations of
on-off control, let us apply the technique
to driving an automobile in a lane of the
road that is marked with white lines. We
know from experience that an auto tends
to go from side to side (to hunt), and
requires continuous steering control.
Let us assume an unreal situation in
which we turn the steering wheel a
small, fixed amount to make the correc-
tion, and do this only when a front wheel
touches a white line. The experiment
would then be like the walk of a blind
person. When crawling along at a very
slow speed we would find that the car
does not go very much outside the lane,
but when driving at a moderate speed
we would find that this type of correction
(applying a fixed amount of adjustment
as on-off control) causes the car to weave
substantially in and out of the lane. If
we were to drive even faster, the car
would be likely to leave the road alto-
gether. The amount of overshoot would
depend on how slowly we respond to vis-
ual signals and take action (see Fig. 6.7).
26
Systems, Feedback, Cybernetics
^\
-&*<
A-
N v%/ X. /") S
y
17"
'i-*£v=
Fig. 6.7. Difficulty of driving an auto by on-off control technique.
Fortunately not many people drive in
this manner because control of an auto-
mobile utilizes a much more sophisticated
system of elements than is possible with
on-off control. In fact, not many automatic
industrial processes can compare with
the sophistication and effectiveness of
good auto driving, since human judgment
enters this operation to a remarkable
degree. To begin with, as the auto moves
to a new position or direction, the driver
is kept continually informed of the nature
of each new situation through his sense of
sight and general physical awareness.
That is, there is continuous feedback, or
information, reaching him to guide his
next move. The element of judgment or
experience also enters. He can vary the
sharpness of turn of the steering wheel
to conform to the sharp right turn. This
is called proportional control. In addition,
he can see a curve in the road ahead long
before the auto has reached the curve.
He can therefore anticipate the move
(anticipatory control) and thus reduce
delay in his action (Fig. 6.8).
The driver of an automobile is aware
of several elements that make control
more difficult. If the steering wheel has
looseness or "play" in the shaft or gear
system, the steering wheel must be turned
several degrees of angle before there is
any effect on the front wheel directions.
This play, or region of no response, is
sometimes called the dead zone of the
system. The driver himself may be a little
slow in judging the situation and taking
action. This "lag or slowness of response
together with looseness in the steering
system, can make for wider overshoot in
the movement of the car. If the throttle
sticks, the motor hesitates, or the brakes
seize, the driver will not be able to assure
smooth "feel" and ride. Finally, rough-
ness of the road can introduce random
fluctuations that add uncertainty to the
normal small feedback of information.
A driver is not likely to give delicate
27
"8
o° ,
c
Q) 3
c -2
o p O
■e — »
< 8„
3 a>
c
"— £ ""
CO
o
1
t
&
w o 4_
CO 0)
o E
c
£ 8? E
8 S 2
" Q) O
O «3
TO O
a> co
0)
CO
a)
** o
o *-
"5 -2 ~
"O — >-
>
" s 5
r
o
*_» z; to
"o c x
Q.
■»- 0 9
c 0
0)
"o
CO
8 ^ TO
2 =5 "5
3 CO <U
M
2 0
_C0
(0
c — ■>
00 _*
«J
< w E
00 c
to <->
a
11
,
1 1
1 i
I 11
<B
CO Ul
CO w
a nee
ces
itrols
ntrol:
o
5 c
5 0
■° c £ o
3
co 2
i; <l> ° o
a
a> a
3 3 ° h_
enn? O o
o
-O Q.
> O
Other di
or in
on aut
peration
3
a.
CO
O H-
Q. °
O O
3 g
° 2
o
T>
_a>
">
o
o
k.
C 3
co
a.
a) .E
3
® 0
2 £
■=:
'
'
•£ > <d to _ >
c «J <u £ -o Q. 2J
c a> tn a> ^ c 9.
.2! o - E c J 8
E
cV
>
o
E
0
2 =»
a> co
E £ c
0 2 «
Is S
>
2 3 *-* O 00 ^ ~
1? 8 Sw>2 ro.S
U
CO
E
T3
E 0 ■* .SP
to co ro CO
0) £ TJ
> CD
c a)
C
TO
0 *-
0
v""' *•- co
ill
Ii
i
1
T3
(A
2
"a
c
a
00 P
c a
*> c
a. c1
3 0JC
S c
3 "C
co 2
CO -o
h, t:
_- 5P
TO '"> .
> ° 2
«J 0) o
28
Systems, Feedback, Cybernetics
guidance to the auto when his whole
body is being shaken. This background
confusion is often called noise or static
when one is referring to transmission of
signal or of information. It exists in almost
every type of control circuit, sometimes
in the form of vibration of an automobile
or plant equipment. It occurs in the
normal radioactivity background of the
environment, which disturbs radiation
measurements. In very sensitive elec-
tronic circuitry it shows up in the random
movements of the electrons. Similar
phenomena are present in social situ-
ations and in biological organisms that
maintain balance in their internal func-
tions and with their environment.
Because cyclic and control aspects of
nature are exceedingly important, we
must consider control principles and no-
menclature in a little more detail before
looking at the several types of systems
that are common.
6.8 Some control principles —
nomenclature
Before beginning detailed discussion we
need some convenient terminology and
symbols for representing the elements
and functions that make up systems.
When the driver of an automobile engages
the gears and steps on the accelerator
pedal, the motor raeexumd the car moves
forward with an expenditure of energy
that is vastly greater than the energy ap-
plied to the pedal. The power is ampli-
fied. We may represent this by a diagram
such as Fig. 6.4. The input, AI in this case,
appears to be simply the change in posi-
tion of the pedal and the small energy
required to make the change. The box in
Fig. 6.4 represents the change or trans-
formation (ol function) that the input
A/ initiates or experiences; in this case
the function produces motor power at a
level that is related (and perhaps propor-
tional) to the position of the throttle or
accelerator pedal. We can refer to the box
as representing a transfer function /(/),
or converter, which produces AO.
What is the source of the energy that
makes this conversion possible? In this
case, the energy source is the chemical
energy in the tank of gasoline. The pedal,
therefore, is nothing more than a lever
device for controlling the use of this
chemical energy. When one includes the
tank of gasoline and the driver along with
the automobile, the system becomes a
closed (or conservative) system.} With-
out either one, the system would be in-
complete. (It is common practice to omit
the sources of energy from block diagrams
of control systems and to indicate only
the energy input and output for a system.)
The operator has freedom to depress
the accelerator pedal quickly or slowly,
as he wishes. A question that is frequently
important for analyzing the behavior of
control systems is the following: What is
the nature of the output response wher
the input is given a quick change? A
quick increment of input change, which
we may represent by A7, is usually called
an input step function.t Figure 6.9 illu-
strates what might happen. Usually there
is some lag in the rise of motor power, as
shown by the curved rise and fall of the
output. This lag will not be serious in
the case of the automobile, since the
input is not likely to be reversed rapidly
very often. In general it is desirable to
have as much lag as one can tolerate,
consistent with adequate control; other-
wise the system will be too ready to
"jump" and probably to overshoot the
f We neglect the fact that as gasoline is
used up it must be replaced, bringing the
entire petroleum industry into our system.
Likewise, food for the driver is neglected.
\ The term step function is often associated
with on-off changes because the change of
power or direction assumes the form of a sud-
den change. This is illustrated by the shape
of the heat input as it is turned on and off in
Figs. 6.9 and 6.10.
29
L-iX
M
Sudden increase of input
T
Sudden Decrease of input
Slow increase of output power level
Slow decrease of output power level
INCREASING TIME-
Fig. 6.9. When the input is given a quick change (step function), the response
of the output may be designed to be slow or rapid. In general, a slow response
of output produces less hunting than does rapid response.
mark and hunt badly before settling down.
The lower curve of Fig. 6.9 illustrates
the nature of the "hunting" that results
when the system is made to respond too
quickly to change in input. How the out-
put will respond depends on the charac-
teristics of the system and on the features
incorporated in the transform function
box of Fig. 6.4.
If the system of moving parts includes
large, heavy components such as the
flywheel and other parts of the automobi le
motor, we can appreciate that quicker
response is possible only if the motor is
designed to have adequate extra power
to give the desired acceleration. But
excessive power can make control less
smooth and more "jumpy," not to mention
excessively costly in gasoline and in the
complexity of the motor itself. The goal
for design of most systems is to find a
happy compromise that makes the system
adequately responsive and yet stable
against excessive hunting, and which is
not too expensive in dollars or in use of
energy.
The system we have been discussing
has the features of proportional control.
That is, the accelerator pedal may be
depressed to give large or small change,
and the motor power level will respond
with some proportional relationship. We
backtrack a little to discuss on-off control
before proceeding further.
6.9 More on on-off control
Earlier we discussed how difficult it
would be to hold an automobile within
the lane of the road if we applied on-off
control principles to adjust the steering.
Despite certain limitations, on-off control
30
Systems, Feedback, Cybernetics
devices are used very commonly in homes
and in industry because of their sim-
plicity. It is very easy to design an elec-
tric iron, an oven, or room-temperature
control, to operate an electric switch
to turn on (or off) the electric power
whenever the temperature falls below
(or rises above) set values. Figure 6.10
illustrates how this might apply to the
thermostat controls for heating a room in
the winter time.
As shown in Fig. 6.10(a), 72°F is the
temperature desired for this room. But
all thermostats and switching devices
require a differential zone of temperature
change in which to go on and off; other-
wise they would act too frequently and
probably erratically because of vibration
conditions and momentary temperature
fluctuations in the immediate neighbor-
hood of the thermostat. We start with the
temperature dropping in the upper curve
of this figure. When the temperature
reaches the lower edge of the differential
B 73'
8. 72c
E
$ 7r
Thermostat goes "off"
to cut off heat
Max room temperature
Thermostat
goes on to ca
for heat
w "off -on" zone
Actual range of
L temperature
change in room
Increasing time
Heater turns "off"
\7I
Heater turns "on"
Increasing time
(a)
r
Above desired level
Thermostat Warm air
Desired temperature level
\
Cool air
Radiator
Below desired level
Fig. 6.10. Relation of room
temperature and thermostat to
the power input to a room-
(b) temperature control system.
31
temperature zone (71°F in Fig. 6.10(a)),
the thermostat switch turns on the heater.
This assumes that there are no significant
time delays in the response of the thermo-
stat or the heater controls. (In actual
experience there are always some delays.)
The radiators around the room take much
more time to heat up, and the temperature
of the air in the room continues to drop
until it reaches some point which is well
below the lower limit of the control
range (about 70° in Fig. 6.10(a)).
As the hot radiators heat the air in the
room, the temperature at the thermostat
starts to climb again, and at the 73°F
level the heaters are turned off. But at that
point the radiators are fully hot, and the
air in the room continues to receive heat
and to rise to a maximum temperature
which is well above 73°F. The net result
is that the room temperature may vary
by as much as four or more degrees
Fahrenheit. In an actual system there will
be a little time lag between the temper-
ature at 71°F or 73°F and the response of
the thermostat and heater controls, which
can make the overshoot and hunting more
severe. Nevertheless, the simplicity and
relatively low cost of on-off systems makes
them very attractive for use in such oper-
ations as controlling temperatures, main-
taining water level in tanks, and many
other operations. Biological and some
social systems, as well as many industrial,
mechanical, and chemical processes,
usually require the more accurate control
that can be achieved through propor-
tional-type systems.
How much power can an on-off system
control? It is fairly clear that the switch
that turns the heater on and off can be
designed to handle any amount of elec-
tric or other form of energy. The amount
depends on the power requirements to
keep the variable that is being controlled
as close to the desired value as possible.
A general rule might be to design the
power level so that the controller calls
for heat about half the time, and the heater
remains off half the total time. Sometimes
the control is improved by supplying a
portion of the power continuously at a
low, fixed level, and allowing the control
system to add or subtract a smaller in-
crement of power as needed.
6.10 Characteristics of
proportional control
The on-off type of temperature control,
in which the power is usually turned
full-on or full-off, is inadequate for many
applications that cannot tolerate the wide
surges around the desired control point
that often accompany on-off systems. The
undesirable surges can be reduced if
the power is moderated in proportion
to the need. This is exactly what is
achieved in proportional control systems,
in which the heat input continues at
some intermediate level when the tem-
perature is near the desired control
point. As the temperature rises somewhat,
the controller reduces the heat input in
proportion to the departure from the set
control point. Similarly, the heat input is
increased in proportion to a fall in tem-
perature below the set control point. Of
course the system becomes more com-
plicated because now the temperature
detector must measure the magnitude of
departure from the control point. (In
on-off control, all that the detector has to
do is to note that the temperature is above
or below the set point.) Also, there must
be somewhat more complex intercon-
nection so that the proportionate (or step-
bv-step) changes in the temperature
detector can be translated into propor-
tionate (or step-by-step) action on the
part of the valve or motor that controls
the fuel or power input.
32
Systems, Feedback, Cybernetics
* 74°-
72°
70°--
Temperature of room near thermostat
Time *-
Response of proportional thermostat to room temperature
Time
<u c
TO °o
52
Heater power level in response to changes in temperature
Time
Fig. 6.11. In a proportional control system, the response of the thermostat is
proportional to the departure of room temperature from the desired control point
and the change in power input to the boiler is proportional to the response of
the thermostat.
Let us analyze the action of such a
system designed to control the temper-
ature in a room.
When the door of the room opens and
lets in a draft of cold air, the thermostat
responds as shown by the drop in the
upper curve of Fig. 6.11. As shown by
the middle curve, at that same time the
thermostat control calls for a propor-
tional increase of heat, and the heaters
respond as shown by the lower curve. As
the draft of cold air becomes warmed
somewhat by mixing with the warmer
air, the proportional thermostat cor-
respondingly reduces its demand for
heat. The net result is that the room tem-
perature is maintained much more closely
to the desired 72°F than is possible with
on-off control. But the proportional control
instruments and equipment tend to be
more expensive, and for that reason they
are not used where on-off control is
adequate.
A serious limitation develops in pro-
portional control systems when the load
demand changes so that a different
average power level must be applied to
hold the variable at the desired control
value. To understand this, we note that
in proportional control, the output AO
(Fig. 6.4) has a fixed ratio to the input
A/. This proportionality ratio, or gain,
may be represented by G = AO/A/. As-
sume that the room-temperature control
we have been discussing is set to control
at 72°F when the outdoor temperature is
around 50°F. We may assume that this
requires an average heat input of 10,000
Btu per hour. Suppose that the outdoor
temperature drops to 0°F. Obviously, the
heater system must provide a great deal
more heat to hold the temperature at
33
72°F, say, 30,000 Btu per hour. We there-
fore need an additional 30,000 - 10,000 =
20,000 Btu per hour to hold the temper-
ature at 72°F. But since in proportional
control more heat is provided only in pro-
portion to the temperature drop from the
control setting, how can the additional
heat be provided without the actual tem-
perature remaining well below the de-
sired control value?
Let us analyze the situation a little
more quantitatively. Suppose that the
gain of our control is set so that, for each
degree that the temperature drops,
the controller permits an additional 2000
Btu per hour to be supplied to the boiler.
This represents a gain or proportionality
ratio of 2000 Btu per hour per degree
fahrenheit. To get the additional 20,000
Btu would require that the temperature
of the room go down to 62°F. Or, alter-
natively, the thermostat setting would
have to be moved arbitrarily to about
80°F in order to supply enough heat to
hold the room temperature at 72°F as long
as the outdoor temperature remained at
zero.
This discrepancy could be reduced if
the gain were made higher (that is, 1°F
could turn on much more than an addi-
tional 2000 Btu per hour). But making the
gain higher also makes the system more
unstable. Other devices can be intro-
duced to change the responsiveness of
the controller, such as incorporating into
the system an outdoor thermostat that
introduces this equivalent of the arbitrary
shift of a thermostat setting. of 80°F. We
need not go into more detail beyond
recognizing this severe limitation of
proportional control systems.
6.11 Feedback
We must give a little more attention to
the important feedback function. When
the thermostat of a temperature-control
system demands more heat, the addi-
tional heat energy continues to pour into
the heater boilers and radiators until
feedback information (in the form of
rising air temperature in its neighbor-
hood) reverses the thermostat demand.
In the case of the driver of the automobile,
although his foot on the accelerator finds
good proportional power response on the
part of the motor, only feedback in the
form of vision (and the transformation of
that ii/ormation into suitable muscle
action) makes driving successful. Without
the presence of feedback, the driver
could not function as part of the system.
The kind as well as the timing (or
phase relationship; see Sect. 6.3) of feed-
back are rather important. In the case of
the temperature controller, the electrical
thermostat reactions must become trans-
formed into heat energy and transfer of
this energy to the room if there is to be
control of temperature. In the case of the
driver, the feedback which arrives in
the form of sensory information must be-
come interpreted and converted into
suitable muscle action on the accelerator
pedal to be effective.
In the case of temperature control the
feedback must always be negative. That
is, the rising room temperature causes the
thermostat to demand less heat, while a
dropping temperature causes it to de-
mand more heat. In the case of driving
an automobile, the feedback may be
negative (say, when the traffic light turns
red and the driver has to let up on the
accelerator) or positive (say, when the
way is clear for higher speed). When a
politician confronts his voting constitu-
ents on an important issue, he watches
their reactions as he talks, to get some
form of feedback, When the response
(or feedback) from the audience is "posi-
34
Systems, Feedback, Cybernetics
tive," he believes that his statements
have been received favorably, whereas
a "negative" feedback is likely to make
him cautious.
Feedback may take many forms and
many types of coupling. Figure 6.12 il-
lustrates a simple modification of an
earlier graph. In this illustration some of
the output energy is fed back to the input.
The box marked "feedback transfer"
determines how much of and in what
form the output will be fed back. The
input is represented by a long arrow with
positive increment. The feedback is
shown as a small arrow with negative
value. In such a setup the net input is
reduced by the amount of the negative
feedback. The effect is to restrain or to
limit the output. If the sign of the feed-
back were positive, the input and the
feedback would add and the output
would increase continually and build
up to destruction, or to the limit of the
energy input. A system with feedback is
often referred to as a closed-loop system.
Since such systems incorporate a measure
of self-correction, the exact value that
the input is permitted to have becomes
less critical. This self-correction factor
also applies to the automobile driver, who
does not have to have a gauge on the foot
pedal because the "feedback" of his eyes
and ears is enough to guide and restrain
his push on the foot pedal.
High values for gain in amplifiers or
control circuits tend to make a system
unstable, and time lags produce wider
oscillations. Negative feedback, on the
other hand, tends to stabilize the systems.
6.12 The elements of
control systems
Now that we have developed some famil-
iarity with control systems, we can iden-
tify the functional elements that make up
most systems.
THE VARIABLE TO BE CONTROLLED
First there is the variable that the
system is expected to cope with or to
control within prescribed limits. Actually
it is rare that only one variable is present
in a system. In the case of room-temper-
ature control the changes in the outdoor
temperature constitute an independent
variable, while the internal temperature
represents the controlled variable. Other
independent variables may be intro-
duced, such as children running in and
A70 =
input +-*-
Converter, amplifier,
transducer, or transfer
function /(/)
+"*~ A0 = OUtput
,
Fig. 6.12.
Addition of a
feedback transfer
function to the
i
'
Feedback
transfer function
transform function
of Fig. 6.4.
35
out of open doors, to cause variable
demands for more or less heating.
Similarly, the driver of the automobile
has control devices by which he steers
and starts and stops the car in relation
to the road. But all along the way he is
forced to comply with independent
demands, such as changing road and
traffic conditions, stop signs, and traffic
lights, all of which constitute independent
variables.
SENSOR DEVICES
Usually there must be some sensor
device by which the variable can be
measured or gauged. For example, in the
case of the temperature measurement we
shall learn in Chapter 10 that the tem-
perature of air is actually determined by
the velocity of the molecules that make
up air. But we cannot gauge the temper-
ature by measuring the velocity of mol-
ecules directly. What we can do is to
utilize some effect that changes with
changing molecule velocity. For example:
At high air temperature, molecules in a
material become more active and bump
each other harder and more frequently,
causing objects such as fluid in a ther-
mometer or a piece of metal in a thermo-
stat to warm up. A thermostat usually
includes a bimetal t that carries an elec-
trical contact; the bimetal changes its
position when the air temperature
changes and thus makes or breaks an
electric circuit.
Similarly, while the position and be-
havior of the automobile are the variables
t A bimetal strip is made up of two different
metals bonded together. Because the two
metals have different temperature expansion
coefficients, the bimetal will bend when
heated, thus causing the contact to switch on
the system. As it cools, it straightens and con-
tact is terminated.
to be controlled, we gauge these by the
use of sensory information (vision, hear-
ing) and the interpretive processes of
the brain. The economist also looks for
meaningful indices by which to gauge
the larger features of national product,
industrial trends, and public attitudes.
The public utilizes quality and creativity
as gauges to evaluate intrinsic or extrinsic-
return on investment.
ENERGY SOURCE
Whether one deals with a temperature-
control system, driving an automobile, or
any other situation that involves variables
and controls, there must be a source of
energy by which the job is performed.
MOTOR TRANSFORM DEVICE
In most instances the sensor function
must utilize the services of a motor device
to restore a variable to its proper value.
To do this the motor device, or motor
function, utilizes energy from an energy
source. In the temperature-control sys-
tem, the blowers and burners (which are
triggered into action by the thermostat)
begin to utilize fuel energy to heat the
boilers. In the automobile a number of
mechanisms come into play to burn the
gasoline, to power the steering, and to
perform other nondriver functions.
FEEDBACK
Finally there is a feedback device, or
feedback function, which in one way or
another relates the output to the input
and thus controls the net output.
The functional elements of a system
cannot always be identified individually
or even as subsystems, but they are
present in one form or another. One
characteristic that will be evident more
and more is the wide variety of trans-
formations (transform functions) that are
36
Systems, Feedback, Cybernetics
possible in systems. Molecular speeds
are transformed into mechanical bending
of a bimetal, which completes an electric
circuit and utilizes electric energy. This
in turn starts a motor and pump device
to feed and burn oil in a boiler, which
produces heat that is transported or trans-
ferred by various means to another area
by other motor devices. Similarly, all
the tangible and intangible features of
human physical energy and human brain
processes become involved with elec-
tromechanical and chemical systems in
driving an automobile. (For each of these
transformations we can apply the more
elegant phrase transform function, and
illustrate the nature of the transformation
by means of a mathematical equation, a
graph, a listing of data, or a simple
picture.)
6.13 Control concepts: cybernetics
The art and science of control theory has
had a long and slow history- In the early
days it found application in the sailing
and steering of ships. With the coming of
the steam engine a mechanical governor
was needed to keep the speed of the
engine constant. In more recent decades
a wide variety of instruments, valves, and
other equipment have been developed
to maintain uniformity in chemical pro-
duction processes. Servomechanisms
were introduced during World War I for
control of gunfire. Electric circuitry and
electromechanical systems were given in-
tensive study to improve their respon-
siveness and stability for purposes of
controlling high-speed operations. By
the 1940's the pace of automation had
quickened as the concepts of control
theory and of feedback received wider
application in the electrical, mechanical,
and processing industries. The term
high-fidelity became a byword in ampli-
fier design as a result of the introduction
of negative feedback.
But the concept of feedback seemed to
be basic and useful for a much wider
range of applications. In 1947 the mathe-
matician Norbert Wiener and Arturo
Rosenbleuth compared the phenomena
of control and of feedback, as used in
technology, to the nervous system and
muscle behavior of the human body.
They postulated a close coordination of
communication relationships between
the brain, the sensory organs, and the
muscles, and concluded that this resulted
from the extensive use of feedback
principles. It seemed that a feedback
function is responsible for one's ability
to reach down and pick up an object and
to know how much farther the hand must
move to complete the act. Moreover, they
found an identity between feedback and
information and the information content
of a signal above the noise level. They
gave the name cybernetics (from the
Greek kybernes for steersman) to the
entire field of control and communication
theory, whether in the machine or in the
animal, f
The concepts oi feedback and informa-
tion encompassed by cybernetics permit
very extensive applications to the bio-
logical and social world. Just as the driver
of the automobile performs functions in
response to the information he derives
t The broad concepts that make up the sci-
ence of cybernetics as developed by Wiener
and his associates were new. The word itself
had much older origin, however. It appears
that Plato often employed the word "cyber-
netics" to mean "the steermans art." His
comment in "Cleitophon," "the cybernetics
of men, as you, Socrates, often call politics,"
suggests a wider implication. In 1834 the
French physicist Ampere used the word as
"means of governing" people.
37
from seeing and hearing and evaluating
the driving situation, so his reactions
under other situations are the result of
his relationship or interaction with each
new environment. Information and feed-
back are essential to his every move,
every decision, almost every thought and
learning process.
We shall have many opportunities to
refer to the principles that have just
been introduced. There will be applica-
tions to strictly technical systems, to
systems that involve nature's resources,
to biological systems, and to social
situations.
The importance of the subject suggests
that we summarize a few of the ideas
that are most pertinent to our purposes.
1. Nature's processes are characterized
by continuous dynamic transformations
of energy, which may range from the
vast magnitudes of astrophysics to the
metabolic adaptation of the smallest
living organism to its environment.
2. Much of man's own activities also
involves the development of processes
for conversion and utilization of nature's
energy resources for purposes of assuring
his survival and comfort. Indeed, the
design of systems that integrate physical
and chemical variables into cooperative,
controlled systems constitutes a main
interest of science and industry to bring
about modern civilization and the current
standard of living of advanced nations.
3. It is now recognized that the ele-
ments that make up a controlled system
have common characteristics, whether
accomplished by machine components,
biological elements, thought processes,
or social situations.
4. In such systems, the element of
feedback, or information, which inter-
relates the output (or behavior) of the
system and the input variables, con-
stitutes a major factor for the effective
operation and stability of systems.
5. The design of every control system
requires careful analysis (and usually
compromise) to meet the needs of the
process. A prime requisite for most con-
trol systems is that they be adequately
responsive to changes, and that they be
stable. Also needed is an adequate source
of energy to perform all the functions
that are required of the system. The input
to the system may be some variable such
as temperature, liquid level, or pressure.
Or it may be information that is itself the
product of other operations, such as in
computer systems.
6. The system performs its function bv
transforming the input to produce an out-
put whose energy content is usually
amplified, the added energy being de-
rived from the source of energy of the
system. The character of the transforma-
tion is designed into the system and is
identified by its transform function to
give the change or amplification gain to
the output.
7. When a feedback (or information)
loop permits some of the energy of the
output to be fed back to the input, there
can be considerable influence on the
nature of the net output and on the
stability of the system. In general, feed-
back that opposes changes (negative) in
the input will improve the stability of
the system, while feedback that arrives
at the input in a manner that increases
its changes (positive feedback) tends to
reduce the stability of a system.
8. The stability of the system suffers
and the system "hunts" more violently
when the amplification or gain between
output and input is too high or when the
system responds too quickly to changes
in the input variable. The design must
include enough damping to reduce exces-
sive overshoot (or violent hunting) of
38
Systems, Feedback, Cybernetics
o
Energy (food)
Brain
Energy losses, work output
Physical
environment
and sensory
system
Physical
man
Information
■ ■ 1 1 ■
Social
Information feedback
or other influence
environment
Fig. 6.13. The
complex inter-
relationship of
man with his
environment.
the system while still providing adequate
response. On-off controls offer cost ad-
vantages and simplicity, but the need for
better control may dictate the use of
proportional control or of other controls
that have more sophisticated design.
There can be more than one input to a
system, more than one output, and a wide
variety of interrelated combinations. In
fact, the input may be the statistical
output of many interrelated elements or
variables.
6.14 Some examples of systems
In Fig. 6.13, which illustrates the rela-
tionship of man to his environment, we
have identified two aspects of man (his
brain and sensory motor system as dis-
tinguished from his physical being) and
two aspects of his enviornment (the
physical and social environments). There
is very intimate and extensive inter-
change between the two aspects of man
and between the two aspects of environ-
ment, as shown by the proximity and
multiple arrows connecting them. Man
draws energy and material from the
physical environment and returns in-
formation and other materials to both.f
In the case of the driving of an automo-
bile, it is difficult to identify all the ele-
ments that make up the input to this
system. The desire to drive, the sensory
activity that provides data to the brain,
and the muscle behavior that operates
the controls of the car, each is a complex
that includes and combines the product
of some other part of the system. The
energy involved in the seeing, hearing,
and judgement operations is negligibly
small, but these become greatly magni-
fied by the body's metabolic processes.
This transform function of the body is
most complex, and is itself made up of
innumerable subsystems.
The specific control principles and
systems we have discussed thus far are
given broader significance by the princi-
ples of cybernetics. Cybernetics deals
f It is not easy to distinguish work from
information and learning. Physical acts are
not readily distinguishable as being separate
from sensory response and interpretation that
leads to learning, judgment, and decision.
Certainly we cannot say that the throwing of
a ball, intake of food, reading of newsprint,
and a walk around the block are not so much
mental processes that lead to future decision
or action as they are physical acts.
39
with elements or variables that are related
to each other so intimately that a change in
one variable is likely to affect other vari-
ables in the system. The elements may
be parts of a machine or those of a chem-
ical process. Cybernetics can deal with
the very specific behavior of a single
molecule among vast numbers of gas
molecules or with the behavior of a
single cell of the vast numbers that make
up an organism. It can as readily (and in
general more usefully) consider the
statistical-behavior character of all the
gas molecules together, or all the cells
of the human body. It can provide a
method for analyzing the economic
relationship of a grocer and his customer,
or as readily attack questions pertaining
to the economics of a whole nation. It
establishes functional relationships in
the course of changes, emphasizing their
coordination, regulation, and control
within a systems concept.
From the point of view of cybernetics,
the aspect of systems behavior that is of
greatest interest is the system's response
to a disturbance. This disturbance may be
a normal change or a momentary depar-
ture (transient) of the input, say as a result
of the dropping of temperature of an in-
dustrial oven below its control setting
when cold material is poured into it. One
or more of the input variables or signals
may experience changes that sum up to
a signal sufficiently large to initiate a
major change. For example, many chem-
ical processes go on within the body,
such as food intake, digestion, blood cell
production, and oxygen utilization. They
are not unrelated and all must be con-
sidered contributory to whether a person
feels well or feels ill. Each process experi-
ences its own daily or hourly variations,
which nevertheless may constitute
normal operation and good health. There
can be occasions, however, when the
individual variations in the processes
add up to produce sickness of a sort that
represents serious imbalance or dis-
turbance of the total system.!
In general, systems are designed to
accept and to cope with very specific
variables and to effect reasonably quick
restoration whenever some change in
those variables upsets equilibrium. The
system is considered to be responsive
when it reacts with adequate speed to the
upset. A system that responds too quickly
or introduces corrective steps that are
too large is likely to produce instability
around the equilibrium point. A system
may also be too sensitive to small fluctu-
ations that are of the order of magnitude
of background "noise," and for that
reason will be unstable.
We might consider the design of an
electrical amplifier system such as that
used for a quality phonograph system.
Figure 6.12 represents a fairly simple
circuit for transforming an input through
some form of transducer to produce an
output. The feedback to the input in this
case was designed to counteract or oppose
the input, tending to reduce undesirable
excursions in the output due to variations
other than the sound signals to be ampli-
fied. The system constitutes a channel
for transmitting and transposing signals,
the input signals being information. To
be effective, the design must usually
incorporate suitable capabilities in such
terms as capacity, watts, voltage, range,
and frequency. These in turn provide the
basis for designing suitable constraints
into the system.
However, a control system is not likely
t As a simple example, the experience of
sitting in an awkward position can introduce
a combination of neural signals and mental
process that suggests the need for a new posi-
tion and thereby requires a complete readjust-
ment of nerve and muscle systems.
40
Systems, Feedback, Cybernetics
to be designed to control every variable
against every change. For example, the
body's control of the iris openings of the
eyes (to permit only adequate light to
reach the retina) has a very specific,
limited function and purpose, which
excludes sensitivity to other variations of
body conditions. The purpose of con-
straints is to reduce the response of the
system to variables that are not con-
sidered to be part of the information to
be transmitted. There are also natural
restraints or constraints on the information
and on the variety of information that a
channel may transmit. Among these are
the limitations and directions imposed
by the conservation of energy and the
laws of thermodynamics. When a system
combines several elements into an inte-
grated organizational and functional
interdependence, the interdependence
automatically imposes constraints, since
the elements are now no longer inde-
pendent of each other. An amplifier
system may have to contend with con-
straints in the form of costs, against which
the designer must balance extra quality
or fidelity or amplification.
With only minor modifications the dia-
gram of Fig. 6.12 can represent a quite
different system for communication of
information. Figure 6.14 illustrates some
Government offices
Editorial and
management
office
Radio and
television
"**] Financial support, etc
Printing
facilities
Distribution
facilities
The
listening
and
reading
public
Financial support, etc.
Fig. 6.14. Schematic representation of a system for communicating news to the
public.
41
of the elements that enter into a system
for communicating news to the general
public. News may be collected from
many areas and reported; this news
becomes input (iu i2, ■ ■ ■ i„) to the
editorial offices of a news agency. At
the editorial offices this information
undergoes modification and shaping, and
is put into printed form or given elec-
trical broadcast. There will be close
liaison among the several blocks that
make up the channels for this communi-
cation. There will be government in-
fluences as well as government sources
of information bearing on the editorial
and management offices, much of it in
the form of feedback reaction to the com-
munication. The listening and reading
public applies "feedback" influence
through financial support (or lack of
support) of the broadcast and publishing
services, through the editorial offices, and
through government offices to the sources
of information. The constraints in such a
system are many. They arise from national
and local government policies; from elec-
trical, chemical, mechanical restraints;
from the cultural habits and educational
level of communities; and from financial
considerations. As a total result, such a
communication system becomes not a
simple amplifier and distributor of simple
news information but also a combination
system for receiving, modifying, trans-
mitting, and generating of news with
built-in restraints and objectives.
One interesting characteristic of a
system of this sort arises from the fact
that any one of the multiple input signals
(h> h, k ■ • ■ in) can suddenly introduce
a major disturbance that overshadows all
other input signals and that can bring
about violent response in either the
forward channels or the feedback chan-
nels. Such a disturbance might be an
act of war, a strike, a catastrophe, or an
event that is especially disliked or espe-
cially desirable. There may be quite a few
surges of output beyond the desired
limits of control before the system set-
tles down again.
One may also conceive that the input
(t, . . . in) can be made up of very many
items and elements so that the overall
significance of the input is determined
by the statistical character of the input
rather than being overly influenced by
any one item.
6.15 Functional relationship:
notations
In its simplest form, a cause-and-effect
relationship is stated as a simple function
such as y =f(x) (meaning y is a function
/ of x, or O =/(/) (meaning output O is
a function of input /). In diagram form
this might be written! as representing a
transition of / into an output O. Relating
this to our earlier example of the auto-
mobile, the power of the motor, O, is
some function of the position of the ac-
celerator pedal, /. If we include the driver
as well, we have a more complete system
with feedback and our equation would
have to provide a different function for
output,
0=f(I)F(0)
In its simplest form the diagram would
be changed to become
' " o
The more interesting examples are not
likely to be so simple as to comprise
t The approach in this section is consider-
ably influenced by the treatment given by
W. Ross Ashbv.
42
Systems, Feedback, Cybernetics
only a single input and single output.
Within the total system that includes the
auto and driver, there are innumerable
smaller systems such as cells, neurons,
muscles, organs, machine parts, and
electrical controls. A study of such as-
semblies must set out clear objectives
before its approach or results can be
made significant. For example, are the
functions to be studied primarily those
that pertain to keeping the auto on the
road, or are they functions that deter-
mine the state of the driver's gall bladder,
heart beat, or temperature? These sub-
sidiary systems are certainly part of
the total system of man and auto, but their
details are independent of the specific
functions that go with driving the auto-
mobile. The situation would be different,
of course, if part of a study had to do with
the effect of heart or temperature func-
tion on the driving, for which purpose a
new set of elements would be involved
when making up the system to be studied.
The situation is illustrated by Ashby
in the following diagram, in which one
may trace twenty different! circuits.
Each subsidiary circuit may have its
own mode of feedback and control, and
may be either strongly or weakly linked
with neighboring circuits. In the case of
our driver, vision plays the dominant
role in telling him where the auto is
going, while his knowledge of the situ-
ation is helped by the senses of hearing
t It is suggested that the reader list the
twenty different ways in which a signal may
travel through the system, starting at A and
returning ultimately to A in each case. For
example, ABCDA or ABCBCDA.
and by the sensations of his body as the
car sways. All the input stimuli have
some relation to each other. One may
picture a strong relationship (or strong
coupling) between vision and hearing
and a weaker coupling between heart-
beat and vision, as far as driving the car
is concerned.
Any study must therefore seek first to
identify the functionally significant rela-
tionships that are the subject of the study,
to identify the elements that bear directly
on the functions under study, and to elimi-
nate from consideration those elements
that are independent of the selected
functions.! This is not easy to do in most
cases because there are many varieties
of influences and "couplings" that come
into play. Often the study must assume a
series of situations and obtain results and
estimates for a wide variety of combina-
tions of systems.
MODELS
Often it becomes necessary to simplify
a system or make it understandable by
use of a model or models. This becomes
imperative for nearly all biological sys-
tems, which are enormously complex.
But models can very quickly become
detrimental to progress when one loses
sight of the simplifications and limita-
tions that are inherent to each model.
6.16 "Black box" approach
Ashby considers the interesting case of
an experimenter approaching a "black
box" that is unknown to him with respect
to contents and functions. How should
he proceed to investigate and determine
the contents and character of the box?
t The importance of such an approach in the
study of human behavior is seen in the Gestalt
psychology view of phenomena.
43
Box
Experimenter
i
7ig. 6.15.
B
lack-box rela
tionship.
(The procedure can be especially impor-
tant if one imagines that the "black box"
could be an explosive bomb.) There is,
immediately, a relationship between the
experimenter and the box, of the type
shown in Fig. 6.15. The diagram illus-
trates the nature of exchanges and feed-
back that take place as the experimenter
explores the problem by various means.
The "means" presumably might include
such things as pushing and pulling of the
box and levers. For a systematic search,
each move would be recorded along with
the "state" of the box that accompanied
each move. In time the experimenter
would presumably be able to identify the
"state" of the box for each type of input,
and possibly also the function for each
type of input. Many of the systems with
which we deal are actually made up of
"black boxes," and the functional charac-
teristics of the total assembly may be de-
termined by the characteristics of each
box and by the nature of their coupling to-
gether, f But we may fail to characterize
the system because a combination of
black boxes may produce an unexpected
function that is quite unrelated to the
characteristics of any one box. An example
given by Ashbv is that the approximately
twenty amino acids in a bacterium do not
individually have the property of being
self-replicating, but their combination
does introduce this property.
Real-life problems tend to have main
"black boxes," often interconnected in
such manner as to obscure the specific
role of each box, each subsystem. One
may make progress in the analysis of the
total system and its parts by systematic
analysis of "responses" or "states" to
questions and input stimuli. One may
seek to discover factors that produce
t The reader may pic tun' the similar situ-
ation that exists when he first meets a person
who is to become liis associate on some project.
44
Systems, Feedback, Cybernetics
certain extreme "responses" of "states."
The use of computers helps handle large
quantities of data and identify common
elements or contrasts. But progress in
attacking complex problems depends
more often on good use of judgment,
experience, intuition or insight, per-
sistence, and some luck. It is not always
easy to identify and isolate the specific
functions that are of importance for the
system's functioning. There may be mul-
titudinous other elements within the
total system that do not bear on the spe-
cific functions under study.
6.17 The closed-loop amplifier
system^
It will be helpful to look a little more
closely at the quantitative aspects of a
system that has feedback characteristics.
Figure 6.16 illustrates a system in which
an input signal E, (which may be in volts
t The sections printed in color represent
optional reading material.
and related to temperature, pressure,
blood count, or other variable) constitutes
the control variable. The system may be
designed to do something that is propor-
tional to or determined by this control
variable i. If the system is a servomech-
anism, input £, may represent the angle
of rotation of a small motor and output
E0 the angle of rotation of a larger motor,
the objective being to keep the two motors
in step with each other. Or £, may be
the input voltage from a measuring circuit
that has high resistance and low power
and which is to be converted to an iden-
tical voltage in a low-resistance circuit
to operate a loudspeaker or solenoid or
some other device that requires more
power than is available at the input end
of the circuit. (Throughout this discus-
sion keep in mind that there must be a
source of energy to make this conversion
possible, as is illustrated in Fig. 6.16.)
The signal £, may have a fixed value
or may vary with time. It feeds into a
comparator element, where E, is com-
pared (added) to the signal coming as
feedback. It both the input signal and
Comparator
E
Input
Input signal which
serves as fixed or
variable point for
control purposes
©
Error or
difference
signal
Ef=(FR)xEc
Energy source
LiiiJ
Converter
amplifier
transformer
effector, etc.
with gain
G
Sensor with
feedback ratio
(FR)
Fig. 6.16. A closed-loop amplifying .system.
GiE.-Ef)
Eo=G(Ei-Ef) + ED
External
disturbance
Output
45
E0 = E
•(t
+ (FR)G
)+Ed\1+(FR)g)
(6-2)
For the above example,
E° = l (l+(lxlO)) + ° (l + 1 x lo)
10 ,
^volt
The error or difference voltage therefore
becomes E, - E, = 1 - (10/11) = 1/11
volt. The amplifier must be capable of
acting on this voltage if the system is de-
signed to work on such magnitudes of
error. One way to improve this is to in-
crease the gain G of the amplifier. If, in
this case, the gain G is increased from 10
to 100 while keeping the feedback ratio
and ED the same, the error voltage reduces
to 1/101 volt from 1/11 volt. It can be seen
that the gain can change markedly with-
out introducing serious error in such a
system.
Finally suppose there develops a dis-
turbance ED amounting to 0.5 volt, with
the gain G = 10 and FR = 1. From the
last term of Eq. (6-2), the effect of the
disturbance reduces to
0.5
1
0.5
1 + 1 x 10 11
as a result of the feedback. Figure 6.17
presents these figures applied directly
to the diagram of Fig. 6.16 (following the
example of James E. Randall, Elements of
Riophysics).
The examples thus far apply to static-
systems. The behavior of systems varies
considerably when the input voltage
changes too rapidly for the system to
follow the changes of E, or ED. The sub-
ject of controller stability has received a
great deal of attention in connection with
servomechanism design (for remote con-
trol of airplane movements, and similar
applications) and electric circuit design
for communication, but we cannot delve
into that aspect of controller theory.
However, one related consideration is
"noise," mentioned briefly in an earlier
discussion. This also has had considerable
study because of the important effect
on the capacity of circuitry • to convey
"information."
6.78 The nature of "living" systems
In the discussions of control systems thus
far we have not distinguished between
systems involving machine components
and those involving living systems. Nor
is it our intention to do so now. The fact
is that, except for varying complexity,
the very same concepts may be applied
to living as well as nonliving or machine
systems. Each of the sensor)' organs
through which we communicate with the
much vaster system of nature is itself
designed, oriented, and functionally
controlled to achieve certain specific
goals or "purposes." It does not matter
whether we discuss a nerve cell or an
electric wire connection. Both are motor-
sensors. Information may be transmitted
through the medium of voice, teletype,
wireless, visual signals, or the raising of
an eyebrow. Each may be an element of a
system, and a composite system may in-
clude many elements or subsystems. The
science of information theory must cope
with vast complications to determine the
maximum and minimum informational
content that an actual system can trans-
mit, even when the role and nature of
each link of the chain can be fairly under-
stood.
The acts of stretching the bodv or of
reaching to pick up an object entail the
function of a fairly complex system of
46
Systems, Feedback, Cybernetics
Comparator
sensitive to
change in sensor
Aldosterone
L Sensor
Na + concentration $ responsive
142 m-eq/l < to Na +
content
-(°).
Adrenal
cortex
<J
production
'
(
JJ
V
ATF
• ADP
\ Intake of
NaCI
Kidney
' /^-—
NaCI loss through
sweating; varies
with activity
Fig. 6.18. A simplified
version of control of
sodium ion concentration
in the extracellular fluid.
Controlled loss
of NaCI
regulation and control. This has been
demonstrated by Karl Smith's! experi-
ments with delayed visual feedback in
visual motor behavior, which showed that
what a person sees is delayed in reaching
him when he is performing various other
tasks such as writing.
The intention to stretch or to pick up an
object is itself a complex function, de-
veloping in the mind as a product of
other activities and influences. The
command signal, in the form of nerve im-
pulses, originate in the motor cortex of
the brain and initiates action in the mus-
cle contractile proteins. There is an
amplification, G, which may be expressed
as change in muscle length for a unit
change in the motor neurons that initiate
the discharges. The muscle spindle acts
as a sensor-transducer to produce nerve
impulses, in proportion to muscle length
extension, to send back to the brain as
f See K. V. Smith.
feedback on the extension. The original
impulses and the feedback impulses are
integrated in the spinal cord and give
indication of the error or difference from
completion of the intended act. The
spindle proprioceptors serve to provide
constant information on the state and tone
of the muscle system, and assure smooth
action of the body. When an individual is
deprived of their help, muscle activity
tends to be abrupt rather than smooth,
requiring dependence on visual sense of
position to the extent that he cannot
stand when blindfolded.
A person suffering from Parkinson's
disease retains some benefit from propri-
oceptive information, but tends to over-
shoot when reaching for an object — a
motion that recalls damped oscillation.
In later studies of biological systems
we shall have occasion to study in some
detail a few of the regulatory systems on
which life depends. Figure 6.18 illus-
trates how the sodium ion concentration
47
is maintained constant in the fluid that
surrounds the individual cells. There is
an elaborate system for maintaining uni-
form pressure in the circulation of blood.
Pressure-sensitive transducers, located in
the aorta and carotid arteries, send infor-
mation about the magnitude of the pres-
sure to an integrating center within the
medullary portion of the brain. This
results in action that lowers blood pres-
sure by slowing the heart rate and also by
producing vascular dilatation.
For respiration there is needed a mini-
mal value for blood carbon dioxide and
an adequate supply of blood oxygen.
When carbon dioxide concentration in
the blood increases, the medullary res-
piratory center stimulates respiration to
eliminate carbon dioxide. The transit
time for the flow of blood between the
lungs and the respiratory center is only a
few seconds under normal conditions (see
Randall, p. 108). Body temperature is
maintained by a delicate balance between
heat loss (from warm-blooded or homeo-
thermic animals) and heat production
within the animal through metabolism.
The "thermostat" that controls this
balance is located in the hypothalmus of
the brain and receives information from
various temperature transducers of the
body to guide its own function.
The regulatory system can extend be-
yond the body to include the interactions
involving climate, geography, geology,
agriculture, theology, government, dis-
ease, or any other influences. The ele-
ments of determinate function, distur-
bances, control variables, amplification,
feedback, informational content, are all
three, but they may take the forms of
imposed law, self-imposed law, self-im-
posed restraints, religious restraints,
moral obligations, and many other forms
that are even less tangible.
The regulatory principles apply to
commercial production plants where
orders for goods become converted to
products for sale, with often quick and
direct feedback from consumer to pro-
ducer. The economist must be aware of
the relationship of the key elements of
a nation's economy in terms that are
identical to those discussed, if he is to
succeed in regulating the ups.and downs
of business within manageable propor-
tions. The problem becomes especially
severe when each of the elements of the
system is a result of statistical variations,
and the statistics lack the assurance of
experience or of numbers. The difficulties
too often savor of the uncertainties of
"black boxes," and yet one must select
a suitable model, suitably simple to be
manageable and not too far removed from
the realities of the situation.
The student is urged to study carefully
all the details that have been included in
this section on controls. In time he will
find that many of the topics that are to
come in later chapters will fall more
easily into place. For nature and man exist
and continue as a result of a balance of
forces and utilization of energy, the whole
constituting a system that is in a state
of reasonable balance and regulation and
yet continually changing toward wholly
new forms.
In conclusion, we hope that this brief
introduction to systems and cybernetics
will encourage each reader to view the
events of his life with keener appreciation
for the interrelationship of the factors
that bear on the events, and especially
for feedback influences. A word of caution
is in order, however, with respect to over-
extended use of the term cybernetics to
situations wherein the relationships are
too complex or too obscure, and wherein
there are not present the control systems
elements which we have discussed.
48
Systems, Feedback, Cybernetics
Questions/Discussions
The assignments for this chapter are
intended to give the reader opportunity
to discover for himself how broadly the
concepts and techniques involving sys-
tems, feedback, control, stability-insta-
bility, and cybernetics apply to phenom-
ena in nature and to all aspects of human
social relations. It is suggested that from
two to four weeks be allowed for comple-
tion of this work.
1. For purposes of review, tabulate the
five elements of control systems (de-
scribed in Sec. 6.12) that apply in the
following personal situations. Explain
also whether the feedback is positive or
negative in each case.
(a) The control of temperature of your
home.
(b) The factors that control your waking
up on a weekday morning.
(c) The factors that control your break-
fasting.
(d) One situation or experience of your
day that includes strong positive
feedback.
(e) A situation or experience of your
day that includes strong negative
feedback.
2. Select three phenomena or situations,
taken from any three of the following
categories, and analyze their "systems"
aspects in the following terms:
(a) The dependent and independent
variables that are involved in each,
either as "input" to the system or
as disturbances.
(b) The sensor devices or transform
functions required at the input end
for each variable.
(c) The energy sources.
(d) The motor devices or processes,
and the related transform functions.
(e) The gain or amplification between
output and input.
(f) The nature of feedback influences
(distinguishing between positive
and negative feedback and phase
relationships) related to each input
and each output.
(g) The nature of subsystems that are
included.
(h) The factors that make for stability
and instability in the total system
or subsystems.
(i) The graphical representation of
the above elements and processes,
with indication of polarity (direc-
tion) of feedback between each out-
put and input.
The phenomena or situations are to be
drawn from any three of the following
seven categories:
I Electromechanical, pneumatic
systems, chemical or production
processes
II Geophysical or meteorological
processes
III Biological processes (plants, ani-
mals), ecological relationships
IV Medical, pathological experiences
V Economics (international, national,
or personal), business operations
VI Behavioral, cultural, ethical, moral,
theological, and psychological
aspects of social experiences
Note: It is suggested that each "case" be
given adequate discussion and one to
two pages of graphical representation.
Because of the importance of the subject
of "systems," it is suggested that these
analyses be given time for class discus-
sion. Group effort on the part of the stu-
dents is encouraged, although each must
present his own final case study.
49
The author, the first American Nobel Prize physicist,
traces the determinations of the velocity of light, one
of the handful of constants of nature.
Velocity of Light
A. A. Michelson
A chapter from his book, Studies in Optics, published in 1927.
The velocity of light is one of the most important of
the fundamental constants of Nature. Its measurement
by Foucault and Fizeau gave as the result a speed greater
in air than in water, thus deciding in favor of the undu-
latory and against the corpuscular theory. Again, the
comparison of the electrostatic and the electromagnetic
units gives as an experimental result a value remarkably
close to the velocity of light — a result which justified
Maxwell in concluding that light is the propagation of an
electromagnetic disturbance. Finally, the principle of
relativity gives the velocity of light a still greater im-
portance, since one of its fundamental postulates is the
constancy of this velocity under all possible condi-
tions.
The first attempt at measurement was due to Galileo.
Two observers, placed at a distance of several kilometers,
are provided with lanterns which can be covered or un-
covered by a movable screen. The first observer uncovers
his light, and the second observer answers by uncovering
his at the instant of perceiving the light from the first.
If there is an interval between the uncovering of the
lantern by the first observer and his perception of the
return signal from the second (due allowance being made
for the delay between perception and motion), the dis-
tance divided by the time interval should give the velocity
of propagation.
51
Needless to say, the time interval was far too small
to be appreciated by such imperfect appliances. It is
nevertheless worthy of note that the principle of the
method is sound, and, with improvements that are almost
intuitive, leads to the well-known method of Fizeau. The
first improvement would clearly be the substitution of a
mirror instead of the second observer. The second would
consist in the substitution of a series of equidistant aper-
tures in a rapidly revolving screen instead of the single
screen which covers and uncovers the light.
The first actual determination of the velocity of light
was made in 1675 by Romer as a result of his observation
of the eclipses of the first satellite of Jupiter. These
eclipses, recurring at very nearly equal intervals, could be
calculated, and Romer found that the observed and the
calculated values showed an annual discrepancy. The
eclipses were later by an interval of sixteen minutes and
twenty-six seconds1 when the earth is farthest from
Jupiter than when nearest to it. Romer correctly attrib-
uted this difference to the time required by light to trav-
erse the earth's orbit. If this be taken as 300,000,000 kilo-
meters and the time interval as one thousand seconds,
the resulting value for the velocity of light is 300,000
kilometers per second.
Another method for the determination of the velocity
of light is due to Bradley, who in 1728 announced an
apparent annual deviation in the direction of the fixed
stars from their mean position, to which he gave the name
"aberration." A star whose direction is at right angles to
the earth's orbital motion appears displaced in the direc-
tion of motion by an angle of 20T445. This displacement
Bradley attributed to the finite velocity of light.
With a telescope pointing in the true direction of such
a star, during the time of passage of the light from ob-
1 The value originally given by Romer, twenty-two minutes, is clearly
too great.
52
Velocity of Light
jective to focus the telescope will have been displaced in
consequence of the orbital motion of the earth so that the
image of the star falls behind the crosshairs. In order to
produce coincidence, the telescope must be inclined for-
ward at such an angle a that the tangent is equal to the
ratio of the velocity v of the earth to the velocity of light,
v
tan a=y ,
or, since v = tD/T, where D is the diameter of the earth's
orbit and T the number of seconds in the year,
tan o-=yf ,
from which the velocity of light may be found; but, as is
also the case with the method of Romer, only to the de-
gree of accuracy with which the sun's distance, §Z>, is
known; that is, with an order of accuracy of about i per
cent.1
In 1849 Fizeau announced the result of the first ex-
perimental measurement of the velocity of light. Two
astronomical telescope objectives Lz and L2 (Fig. 73) are
placed facing each other at the two stations. At the focus
of the first is an intense but minute image a of the source
of light (arc) by reflection from a plane-parallel plate N.
The light from this image is rendered approximately
parallel by the first objective. These parallel rays, falling
on the distant objective, are brought to a focus at the sur-
face of a mirror, whence the path is retraced and an
image formed which coincides with the original image a,
where it is observed by the ocular E. An accurately
divided toothed wheel W is given a uniform rotation,
1 The value of the velocity of light has been obtained, by experimental
methods immediately to be described, with an order of accuracy of one in
one hundred thousand, so that now the process is inverted, and this re-
sult is employed to find the sun's distance.
53
*s3
IN
Fig. 73
thus interrupting the passage of the light at a. If, on
returning, the light is blocked by a tooth, it is eclipsed, to
reappear at a velocity such that the next succeeding
interval occupies the place of the former, and so on.
If n is the number of teeth and N the number of turns
per second, K the number of teeth which pass during the
double journey of the light over the distance D,
V =
2NnD
K
It is easier to mark the minima than the maxima of in-
tensity, and accordingly
2
if p is the order of the eclipse. Let bK be the error com-
mitted in the estimate of K (practically the error in esti-
mation of equality of intensities on the descending and
the ascending branches of the intensity curve). Then
dV = dK
V K '
Hence it is desirable to make K as great as possible. In
Fizeau's experiments this number was 5 to 7, and should
54
Velocity of Light
have given a result correct to about one three-hundredth.
It was, in fact, about 5 per cent too large.
A much more accurate determination was undertaken
by Cornu in 1872 in which K varied from 3 to 21, the re-
sult as given by Cornu being 300,400, with a probable er-
ror of one-tenth of 1 per cent. In discussing Cornu's re-
sults, however, Listing showed that these tended toward a
smaller value as the speed increased, and he assigns this
limit as the correct value, namely, 299,950. Perrotin,
with the same apparatus, found 299,900.
Before Fizeau had concluded his experiments, another
project was proposed by Arago, namely, the utilization
of the revolving mirror by means of which Wheatstone
had measured the speed of propagation of an electric
current. Arago's chief interest in the problem lay in the
possibility of deciding the question of the relative veloci-
ties in air and water as a crucial test between the undula-
tory and the corpuscular theories. He pointed out,
however, the possibility of measuring the absolute ve-
locity.
The plan was to compare the deviations of the light
from an electric spark reflected directly from the revolving
mirror with that which was reflected after traversing a
considerable distance in air (or in water). The difficulty
in executing such an experiment lay in the uncertainty in
the direction in which the two reflected images of the
spark were to appear (which might be anywhere in 3600).
This difficulty was solved by Foucault in 1862 by the
following ingenious device whereby the return light is
always reflected in the same direction (apart from the
deviation due to the retardation which it is required to
measure), notwithstanding the rotation of the mirror.
Following is the actual arrangement of apparatus by
which this is effected. Light from a source S falls upon
an objective L, whence it proceeds to the revolving mir-
ror R, and is thence reflected to the concave mirror C
55
Fig. 74
(whose center is at R), where it forms a real image of the
source. It then retraces its path, forming a real image
which coincides with the source even when the revolving
mirror is in slow motion. Part of the light is reflected from
the plane-parallel glass M , forming an image at a where
it is observed by the micrometer eyepiece E.
If now the revolving mirror is turning rapidly, the
return image, instead of coinciding with its original posi-
tion, will be deviated in the direction of rotation through
an angle double that through which the mirror turns while
the light makes its double transit. If this angle is a and
the distance between mirrors is D, and the revolving
mirror makes N turns per second,
a = 27riv-^- i
or
V =
4irND
In principle there is no essential difference between
the two methods. In the method of the toothed wheel the
angle a corresponds to the passage of K teeth, and is
therefore a = 2irK/n, so that the formula previously found,
V =
2NnD
K
j now becomes V =
4irND
the same as for the
56
Velocity of Light
revolving mirror. The latter method has, however, the
same advantage over the former that the method of mir-
ror and scale has over the direct reading of the needle of
a galvanometer.
On the other hand, an important advantage for the
method of the toothed wheel lies in the circumstance
that the intensity of the return image is one-half of that
which would appear if there were no toothed wheel,
nB
whereas with the revolving mirror this fraction is —
ru
if the mirror has n facets) , where /3 is the angular aperture
of the concave mirror, and / is the focal length of the
mirror, r is the distance from slit to revolving mirror,
and D is the distance between stations.
In the actual experiments of Foucault, the greatest
distance D was only 20 m (obtained by five reflections
from concave mirrors), which, with a speed of five hun-
dred turns per second, gives only 160" for the angle 2a
which is to be measured. The limit of accuracy of the
method is about one second, so that under these condi-
tions the results of Foucault's measurements can hardly
be expected to be accurate to one part in one hundred and
sixty. Foucault's result, 298,000, is in fact too small by
this amount.1
In order to obtain a deflection 2a sufficiently large to
measure with precision it is necessary to work with a
much larger distance. The following plan renders this
possible, and in a series of experiments (1878) the dis-
tance D was about 700 m and could have been made much
greater.
1 Apart from the mere matter of convenience in limiting the distance
D to the insignificant 20 m (on account of the dimensions of the labora-
tory), it may be that this was in fact limited by the relative intensity of
the return image as compared with that of the streak of light caused by
the direct reflection from the revolving mirror, which in Foucault's
experiments was doubtless superposed on the former. The intensity of
the return image varies inversely as the cube of the distance, while
that of the streak remains constant.
57
The image-forming lens in the new arrangement is
placed between the two mirrors, and (for maximum in-
tensity of the return image) at a distance from the re-
volving mirror equal to the focal length of the lens. This
necessitates a lens of long focus; for the radius of meas-
urement r (from which a is determined by the relation
8 = r tan a, in which 8 is the measured displacement of the
image) is given by r = -= , if / is the focal length of the lens ;
whence r is proportional to/2. In the actual experiment,
a non-achromatic lens of 2 5-m focus and 20-cm diameter
was employed, and with this it was found that the in-
tensity of the return light was quite sufficient even when
the revolving mirror was far removed from the principal
focus.
With so large a displacement, the inclined plane-paral-
lel plate in the Foucault arrangement may be suppressed,
the direct (real) image being observed. With 250 to 300
turns per second, a displacement of 100 to 150 mm was
obtained which could be measured with an error of less
than one ten-thousandth.
The measurement of D presents no serious difficulty.
This was accomplished by means of a steel tape whose
coefficient of stretch and of dilatation was carefully deter-
mined, and whose length under standard conditions was
compared with a copy of the standard meter. The esti-
mated probable error was of the order of 1 : 200,000.
The measurement of the speed of rotation presents
some points of interest. The optical "beats" between the
revolving mirror and an electrically maintained tuning
fork were observed at the same time that the coincidence
of the deflected image with the crosshairs of the eyepiece
was maintained by hand regulation of an air blast which
actuated the turbine attached to the revolving mirror.
The number of vibrations of the fork plus the number of
beats per second gives the number of revolutions per
58
Velocity of Light
second in terms of the rate of the fork. This, however,
cannot be relied upon except for a short interval, and
it was compared before and after every measurement
with a standard fork. This fork, whose temperature co-
efficient is well determined, is then compared, as follows,
directly with a free pendulum.
For this purpose the pendulum is connected in series
with a battery and the primary of an induction coil whose
circuit is interrupted by means of a platinum knife edge
attached to the pendulum passing through a globule of
mercury. The secondary of the induction coil sends a
flash through a vacuum tube, thus illuminating the edge
of the fork and the crosshair of the observing microscope.
If the fork makes an exact whole number (256) of vibra-
tions during one swing of the pendulum, it appears at
rest; but if there is a slight excess, the edge of the fork
appears to execute a cycle of displacement at the rate of n
per second. The rate of the fork is then N±n per second
of the free pendulum. This last is finally compared with
a standard astronomical clock.1 The order of accuracy
is estimated as 1 : 200,000.
The final result of the mean of two such determina-
tions of the velocity of light made under somewhat similar
conditions but at a different time and locality is 299,895.
A determination of the velocity of light by a modifica-
tion of the Foucault arrangement was completed by
Newcomb in 1882. One of the essential improvements
consisted in the use of a revolving steel prism with square
section twice as long as wide. This permits the sending
and receiving of the light on different parts of the mirror,
thus eliminating the effect of direct reflection. It should
also be mentioned that very accurate means were pro-
vided for measuring the deflection, and finally that the
1 The average beat of such a clock may be extremely constant al-
though the individual "seconds" vary considerably.
59
speed of the mirror was registered on a chronograph
through a system of gears connected with the revolving
mirror. Newcomb's result is 299,860.
The original purpose of the Foucault arrangement was
the testing of the question of the relative velocities of light
in air and in water. For this purpose a tube filled with
water and closed with plane-parallel glasses is interposed.
There are then two return images of the source which
would be superposed if the velocities were the same. By
appropriately placed diaphragms these two images may
be separated, and if there is any difference in velocities
this is revealed by a relative displacement in the direction
of rotation. This was found greater for the beam which
had passed through the water column, and in which,
therefore, the velocity must have been less. This result
is in accordance with the undulatory theory and opposed
to the corpuscular theory of light.
The experiments of Foucault do not appear to have
shown more than qualitative results, and it should be of
interest, not only to show that the velocity of light is less
in water than in air, but that the ratio of the velocities
is equal to the index of refraction of the liquid. Experi-
ments were accordingly undertaken with water, the result
obtained agreeing very nearly with the index of refrac-
tion. But on replacing the water by carbon disulphide,
the ratio of velocities obtained was 1.75 instead of 1.64,
the index of refraction. The difference is much too great
to be attributed to errors of experiment.
Lord Rayleigh found the following explanation of the
discrepancy. In the method of the toothed wheel the dis-
turbances are propagated in the form of isolated groups
of wave-trains. Rayleigh finds that the velocity of a
group is not the same as that of the separate waves ex-
cept in a medium without dispersion. The simplest form
of group analytically considered is that produced by two
60
Velocity of Light
simple harmonic wave-trains of slightly different fre-
quencies and wave-lengths. Thus, let
y=cos (nt— mx)+cos (nj—nhx) ,
in which n = 2ir/T, and m = 2ir/\, T being the period and
X the wave-length. Let n—nz = dn, and m—mi=dm.
Then
y=2 cos %(dnt— dmx) cos (nt—tnx) .
This represents a series of groups of waves such as illus-
trated in Figure 75.
Fig. 75
The velocity of the waves is the ratio V = n/m, but
the velocity of the group (e.g., the velocity of propagation
of the maximum or the minimum) will be
V' = dn/dm,
or, since n = mV,
or, since w = 27r/X,
\ vd\J
The demonstration is true, not only of this particular
form of group, but (by the Fourier theorem) can be ap-
plied to a group of any form.
It is not quite so clear that this expression applies to
the measurements made with the revolving mirror. Lord
Rayleigh shows that in consequence of the Doppler effect
there is a shortening of the waves at one edge of the
61
beam of light reflected from the revolving mirror and a
lengthening at the opposite edge, and since the velocity
of propagation depends on the wave-length in a dispersive
medium, there will be a rotation of the individual wave-
fronts.
If to is the angular velocity of the mirror, and wx that
of the dispersional rotation,
= dV = dVd\
dy d\ dy '
where y is the distance from the axis of rotation. But
d\ X X dV
Ty^^f'-^^VdX'
The deflection actually observed is therefore
r(2w+a>i) ,
where T is the time required to travel distance 2D; or
/ ,\dV\
\1+Vdk)>
hence the velocity measured is
•"-'♦MS).
or, to small quantities of the second order,
V" =V' = group velocity -1
The value of ( H — -tt ) for carbon disulphide for
the mean wave-length of the visible spectrum is 0.93-.
Accordingly,
V V 0.93 0.93 ' '
which agrees with the value found by experiment.
' J. W. Gibbs {Nature, 1886) shows that the measurement is in reality
exactly that of groups and not merely an approximation.
f-l
62
Velocity of Light
RECENT MEASUREMENTS OF THE VELOCITY OF LIGHT
In the expression for V, the velocity of light as de-
termined by the revolving mirror, V = 4tND/cl, there are
three quantities to be measured, namely, N, the speed of
the mirror; D, the distance between stations; and a, the
angular displacement of the mirror. As has already been
mentioned, the values of N and D may be obtained to
one part in one hundred thousand or less. But a cannot
be measured to this order of accuracy. It has been pointed
out by Newcomb1 that this difficulty may be avoided by
giving the revolving mirror a prismatic form and making
the distance between the two stations so great that the
return light is reflected at the same angle by the next fol-
lowing face of the prism.
The following is an outline of a proposed attempt to
realize such a project between Mount Wilson and Mount
San Antonio near Pasadena, the distance being about
35 km. For this, given a speed of rotation of 1,060 turns
per second, the angular displacement of the mirror during
the double journey would be 900; or, if the speed were
half as great, an angle of 450 would suffice.2 Accordingly,
the revolving mirror may have the form of an octagon.
It is, of course, very important that the angles should be
equal, at least to the order
of accuracy desired.
This has already been
attained as follows. The
octagon, with faces pol-
ished and angles approxi-
mately correct, is applied
to the test angle a V made
up of a 450 prism ce-
mented to a true plane.
The faces btb are made
parallel by the interfer-
ence fringes observed in
Fig. 76
1 Measures of the Velocity of Light. Nautical Almanac Office, 1882.
2 It may be noted that with eight surfaces the resulting intensity will
be four times as great as with the revolving plane-parallel disk.
63
monochromatic light. In general, the faces aLa will not be
parallel, and the angle between them is measured by the
distance and inclination of the interference bands. The
same process is repeated for each of the eight angles, and
these are corrected by repolishing until the distance and
inclination are the same for all, when the corresponding
angles will also be equal. It has been found possible in
this way to produce an octagon in which the average
error was of the order of one-millionth, that is, about
one-tenth to one-twentieth of a second.1
Another difficulty arises from the direct reflection and
the scattered light from the revolving mirror. The former
may be eliminated, as already mentioned, by slightly
inclining the revolving mirror, but to avoid the scattered
b'ght it is essential that the return ray be received on a
different surface from the outgoing.
zee 3
Fig. 77.— Light path a, b, c, d, e, eufu U, e,f, g, h, i,j
1 It may be noted that while a distortion may be expected when the
mirror is in such rapid rotation, if the substance of the mirror (glass, in
the present instance) is uniform, such distortion could only produce a very
slight curvature and hence merely a minute change of focus.
64
Velocity of Light
Again, in order to avoid the difficulty in maintaining
the distant mirror perpendicular to the incident light, the
return of the ray to the home station may be accom-
plished exactly as in the Fizeau experiment, the only pre-
caution required being the very accurate focusing of the
beam on the small plane (better, concave) mirror at the
focus of the distant collimator.
Finally, it is far less expensive to make both sending
and receiving collimators silvered mirrors instead of
lenses.
In Figure 77 is shown the arrangement of apparatus
which fulfilled all these requirements.
Three determinations were undertaken between the
home station at the Mount Wilson Observatory and
Mount San Antonio 22 miles distant. The rate of the
electric tuning fork was 132.25 vibrations per second,
giving four stationary images of the revolving mirror
when this was rotating at the rate of 529 turns per second.
The fork was compared before and after every set of the
observations with a free pendulum whose rate was found
by comparison with an invar pendulum furnished and
rated by the Coast and Geodetic Survey.
The result of eight measurements in 1924 gave
Va= 299,735 .
Another series of observations with a direct compari-
son of the same electric fork with the Coast and Geodetic
Survey pendulum1 was completed in the summer of 1925
with a resulting value
Va = 299,690 .
A third series of measurements was made in which the
electric fork was replaced by a free fork making 528 vibra-
1 This comparison was made by allowing the light from a very narrow
slit to fall on a mirror attached to the pendulum. An image of the slit was
formed by means of a good achromatic lens, in the plane of one edge of the
fork, where it was observed by an ordinary eyepiece.
65
tions per second maintained by an "auction circuit," thus
insuring a much more nearly constant rate. The result
of this measurement gave
Va= 299,704 .
Giving these determinations the weights 1, 2, and 4,
respectively, the result for the velocity in air is
Va= 299,704 .
Applying the correction of 67 km for the reduction
to vacuo gives finally 7=299,771 .
This result should be considered as provisional, and
depends on the value of D, the distance between the two
stations which was furnished by the Coast and Geodetic
Survey, and which it is hoped may be verified by a repeti-
tion of the work.
It was also found that a trial with a much larger
revolving mirror gave better definition, more light, and
steadier speed of rotation; so that it seems probable that
results of much greater accuracy may be obtained in a
future investigation.
FINAL MEASUREMENTS
Observations with the same layout were resumed in
the summer of 1926, but with an assortment of revolving
mirrors.
The first of these was the same small octagonal glass
mirror used in the preceding work. The result obtained
this year was V— 299,813. Giving this a weight 2 and
the result of preceding work weight 1 gives 299,799 for
the weighted mean.
The other mirrors were a steel octagon, a glass 12-
sider, a steel 12-sider, and a glass 16-sider.
66
Velocity of Light
The final results are summarized in Table VII.
TABLE VII
Turns per Second
Mirror
Number of
Observations
Vel. of Light
in Vacuo
528
Glass oct.
Steel oct.
Glass 12
Steel 12
Glass 16
576
195
270
2l8
504
299,797
299,795
299,796
299,796
299,796
528
352
352
264
Weighted mean, 299,796 + 1
Table VIII shows the more reliable results of measure-
ments of V with distance between stations, method used
and the weight assigned to each.
TABLE VIII
Author
D
Method
Wt.
V
Cornu
23 km
12
0.6
6-5
35
Toothed wheel
Toothed wheel
Rev. mirror
Rev. mirror
Rev. mirror
1
1
1
3
5
299,990
299 , 900
299,880
299,81(7
299,800
Perrotin
Mr and M2
Newcomb*
M3
* Newcomb's value omitting all discordant observations was 298,860.
67
Bees, water fleas, and horseshoe crabs navigate by
polarized light. Sunglasses, camera filters, and
glare-free auto headlights are among other applica-
tions.
Popular Applications of Polarized Light
William A. Shurcliff and Stanley S. Ballard
A chapter from the book Polarized Light published in 1964.
If there is a logical order in which the various applications of
polarizers and polarized light should be considered, the authors
have never discovered it. The policy adopted here is to consider
the most popular and "humanistic" applications first, and the
more scientific and esoteric applications last.
POLARIZATION AND THE HUMAN EYE
The most humanistic fact about polarization of light is that
it can be detected directly by the naked eye. Nearly anyone, if
told carefully what to look for, can succeed in this. Sometimes he
can even determine the form and azimuth of polarization.
What the observer actually "sees" is a certain faint pattern
known as Haidinger's brush and illustrated in Fig. 10-1. The
brush is so faint and ill-defined that it will escape notice unless
the field of view is highly uniform: a clear blue sky makes an
ideal background, and a brightly illuminated sheet of white
paper is nearly as good. The best procedure for a beginner is to
hold a linear polarizer in front of his eye, stare fixedly through
it toward a clear blue sky, and, after five or ten seconds, sud-
denly turn the polarizer through 90°. Immediately the brush is
seen. It fades away in two or three seconds, but reappears if the
polarizer is again turned through 90°. The brush itself is sym-
69
BLUE
, '. YELL9wJ§ffl| ^Syellow .* •* • ,
• • BLUE •- -
ABOUT 3 DEGREES
FIG. 10-1 Approximate appearance of Haidinger's brush when the
vibration direction of the beam is vertical.
metric, double-ended, and yellow in color; it is small, subtending
an angle of only about 2° or 3°. The adjacent areas appear blue,
perhaps merely by contrast. The long axis of the brush is approx-
imately perpendicular to the direction of electric vibration in the
linearly polarized beam, i.e., perpendicular to the transmission
axis of the polarizer used.
Circular polarization, too, can be detected directly by eye, and
even the handedness can be determined. When an observer fac-
ing a clear blue sky places a right circular polarizer in front of
his eye, he sees the yellow brush and finds that its long axis has
an upward-to-the-right, downward-to-the-left direction, i.e., an
azimuth of about +45°. This is true, of course, irrespective of the
orientation of the polarizer, since a circle has no top or bottom.
If he employs a left circular polarizer, he finds the brush to have
a —45° orientation. In each case the pattern fades away rapidly,
but can be restored to full vigor by switching to a polarizer of
opposite handedness. Instead of using a circular polarizer the
observer can use a single linear polarizer in series with a 90° re-
tarder, the latter being held nearer to the eye. Turning the
retarder through 90° reverses the handedness of the circular
polarization.
Some people see the brush easily; others have difficulty. A few
70
Popular Applications of Polarized Light
see the brush when looking innocently at the partially polarized
blue sky, i.e., without using any polarizer at all, and even with-
out meaning to see the brush. Some people see the brush more
distinctly by linearly polarized light than by circularly polarized
light, and for others the reverse is true. An observer may find
the brush to have a slightly different orientation depending on
which eye is used.
The spectral energy distribution of the light is important. If
the light is rich in short-wavelength (blue) radiation, the brush
is very noticeable, but if the short-wavelength radiation is elimi-
nated by means of a yellow filter, the brush fails to appear. Use
of a blue filter tends to accentuate the brush.
Although the phenomenon was discovered in 1844, by the
Austrian mineralogist Haidinger, the cause is not yet fully under-
stood. Presumably the thousands of tiny blue-light-absorbing
bodies in the central (foveal) portion of the retina are dichroic
and are oriented in a radial pattern, for example, a pattern such
that the absorption axis of each body lies approximately along
a radius from the center of the fovea. Incident linearly polarized
light will then be absorbed more strongly in some parts of the
pattern than in other parts and consequently some parts will
fatigue more than others. When the vibration direction of the
light is suddenly changed, the varying degrees of fatigue are
revealed as a subjective radial pattern. Presumably no such
dichroism or orientation pattern applies to longer wavelength
(yellow and red) light; consequently a yellow sensation domi-
nates in those regions where fatigue-to-blue has occurred.
The fact that circular polarization, also, may be detected per-
haps implies that some transparent portion of the eye is weakly
birefringent and acts like a retarder, converting circularly polar-
ized light to linearly or elliptically polarized light. The direc-
tion of the major axis of the ellipse depends only on the direc-
tion of the fast axis of the retarding layer and hence remains
fixed — unless the observer tips his head.
Perhaps physicists will some day write matrices to describe the
retarding layers and dichroic areas of the eye. Poets were the first
to see magic fire and jewels in the human eye; physicists will be
the first to see matricesl
71
Bees, too, can detect the vibration direction of linearly polar-
ized light. The experiments of the biologist K. von Frisch during
World War II showed that bees "navigate" back and forth be-
tween hive and source of honey by using the sun as a guide.
More interesting, when the sun is obscured by a large area of
clouds the bees can still navigate successfully if they can see a
bit of blue sky: they can detect the azimuth of linear polariza-
tion of the blue light and navigate with respect to it. One way of
demonstrating the bee's ability to detect the azimuth of polariza-
tion is to place the bee in a large box the top of which consists
of a huge sheet of linear polarizer, such as H-sheet. Each time the
experimenter turns the polarizer to a different azimuth, the bee
changes his direction of attempted travel correspondingly.
Certain other animals also can detect the polarization of sky-
light and navigate by it. This includes ants, beetles, and the
fruit fly Drosophila. Probably many other examples will be dis-
covered.
POLARIZATION OF SKY LIGHT
Blue-sky light traveling in a direction roughly at right angles
to the sun's rays is partially polarized. When an observer holds
a linear polarizer in front of his eye and gazes in a direction
perpendicular to the direction of the sun, he finds that rotating
the polarizer slowly causes the sky to change from bright to dark
successively. The degree of polarization of sky light may reach
70 or 80 percent when the air is clear and dust-free, the sun is
moderately low in the sky, and the observation direction is near
the zenith.
The polarization is a result of the scattering of the sun's rays
by the molecules in the air. Rayleigh's well-known inverse-fourth-
power law relating scattering intensity to wavelength accounts
for the blue color of the scattered light, and the asymmetry as-
sociated with the 90° viewing angle accounts for the polarization,
as explained in Chapter 5. Some multiple scattering occurs, and
this reduces the degree of polarization somewhat; when the
observer ascends to a higher altitude, the amount of air involved
72
Popular Applications of Polarized Light
is reduced, multiple scattering is reduced, and the degree of
polarization is increased. A further increase results when a
yellow or red filter is used to block the short-wavelength com-
ponent of the light and transmit the long-wavelength component
— the latter component is less subject to multiple scattering. (The
situation is very different for infrared radiation of wavelength
exceeding 2 microns: much of this radiation is produced by
emission from the air itself, rather than by scattering, and this
exhibits little or no polarization.)
Some persons are capable of detecting the polarization of sky
light directly by eye, by virtue of the Haidinger brush phe-
nomenon discussed in a preceding section; a few individuals
find the brush noticeable enough to be a nuisance. Ordinarily,
of course, it escapes notice and plays little part in the affairs of
man. Its practical use by bees, ants, etc., has been indicated, and
the importance to photographers is discussed in a later section.
POLARIZATION OF LIGHT UNDER WATER
A surprising fact about the polarization found in light present
beneath the surface of the ocean (or of a pond) is that the pre-
dominant direction of electric vibration is horizontal. The oppo-
site might be expected, since most of the light that enters the
water enters obliquely from above, and the most strongly re-
flected component of obliquely incident light is the horizontally
vibrating component. But oceanographers and biologists, work-
ing at depths of 5 to 30 feet in waters off Bermuda and in the
Mediterranean Sea, have found the main cause of submarine
polarization to be the scattering of the light by microscopic
particles suspended in the water. Sunlight and sky light enter
the water from above, and the average direction of illumination
is roughly vertical; consequently the polarization form of the
scattered light that travels horizontally toward an underwater
observer is partially polarized with the electric vibration direc-
tion horizontal. The situation is much the same as that discussed
in Chapter 5, except that the incident light has a more steeply
downward direction and the asymmetric scattering is by micro-
scopic particles instead of molecules.
Typically, the degree of polarization is 5 to 30 percent, an
73
amount found to be important to a variety of underwater life.
The water flea Daphnia tends to swim in a direction perpendicu-
lar to the electric vibration direction, for reasons not yet known.
When tests are conducted in a tank filled with water that is free
of suspended particles, so that the submarine illumination is
practically unpolarized, Daphnia ceases to favor any one direc-
tion. But if suspended matter is added, thus restoring the polar-
ization, Daphnia resumes the custom of traveling perpendicular
to the vibration direction.
The arthropod Limulus (horse shoe crab) easily detects the
polarization of the underwater light and is presumed to navigate
with respect to the electric vibration direction. The same is true
of the crustacean Mysidium gracile and various other forms of
marine life. Most tend to swim perpendicularly to the vibration
direction; some swim parallel to it; a few swim at different rela-
tive orientations depending on the time of day. For all of these
animals, polarization is a compass that works even under water!
POLARIZING SUNGLASSES
The lenses of ordinary sunglasses employ absorbing materials
that are isotropic, and accordingly the incident light is attenuated
by a fixed factor irrespective of polarization form. This is un-
fortunate. The fact is that "glare" consists predominantly of
light having a horizontal vibration direction. Why? For these
reasons:
(a) The main source of light (sun and sky) is overhead, and
consequently the main flux of light is downward.
(b) The surfaces that are most strongly illuminated by the
downward flux are horizontal surfaces.
(c) Such surfaces are usually viewed obliquely, since a person
seldom looks straight down.
(d) Most outdoor objects are of dielectric material.
(e) Light reflected obliquely from a horizontal dielectric sur-
face is partially linearly polarized with the dominant vibration
direction horizontal, as explained in Chapter 4.
Polarizing sunglasses take full advantage of this fact. The
lenses are made of dichroic material (H-sheet, usually) oriented
with the transmission axis vertical, as indicated in Fig. 10-2a, so
74
Popular Applications of Polarized Light
(o)
(b)
(c)
FIG. 10-2 Three types of polarizing spectacles. In (a) the transmission
axis is vertical, for eliminating glare reflected from horizontal surfaces.
In (b) the axis is horizontal, for eliminating reflections from vertical
windows of trains, store-fronts (show-windows), etc. In (c) the axis di-
rections are 45* and —45°, a standard arrangement used in viewing
polarization-coded stereoscopic pictures.
that almost all of the horizontal vibrations are absorbed. The
component having vertical vibration direction is transmitted.
Usually some isotropic absorber is included in the lenses to
absorb ultraviolet light strongly and blue and red light to a
moderate extent; the sunglasses then have a greenish hue which
has nothing to do with the polarization.
Motorists and vacationists find that polarizing sunglasses are
helpful not only in reducing the brightness of the field of view
as a whole, but also in enhancing the beauty of the scene. Be-
cause specularly reflected light is absorbed preferentially, roads,
trees, grassy fields, etc., appear softer and more deeply colored
through polarizers. Specularly reflected light tends to veil nature's
inherent beauty; polarizing sunglasses remove the veil.
Fishermen and boatsmen enjoy another benefit from wearing
polarizing sunglasses. They want to be able to see fish, rocks, etc.,
beneath the surface of the water, yet the light from such objects
is dim and is usually lost in the "noise" of the sky light reflected
obliquely from the surface. Since the reflected light is highly
polarized with horizontal vibration direction, the polarizing
sunglasses absorb this component strongly, and the visibility of
75
the underwater objects is greatly increased. The increase is great-
est when the viewing direction corresponds to the polarizing
angle, which, for water, is about 53° from the normal. When the
viewing direction is along the normal, i.e., straight down, there
is no increase at all.
There is one interesting situation in which polarizing sun-
glasses produce little increase in visibility of underwater objects
even when the angle of viewing is the polarizing angle. This situ-
ation occurs when the sky is clear and blue, the sun is low in
the sky, and the pertinent portion of the sky is at 90° from the
direction of the sun. Under these circumstances the light striking
the water is already linearly polarized at such an azimuth that
almost none of it is reflected. There is no task left for the sun-
glasses to perform — there is no reflected glare to suppress. The
underwater objects are seen with great clarity. Persons unfamiliar
with the polarization of sky light and with the dependence of
oblique reflection on polarization form are likely to ascribe the
remarkable clarity to "especially clear water" rather than to
absence of reflection.
CAMERA FILTERS
Photographers often wish to enhance the contrast between blue
sky and white clouds. Thirty years ago they did this by employing
a yellow filter, which absorbed most of the blue light from the
clear sky but transmitted most of the white light from the clouds.
Using ordinary black-and-white film, they obtained excellent
contrast by this method. Today, photographers are using color
film increasingly, and the use of yellow filters is no longer per-
missible since it would eliminate all blue colors from the finished
photograph.
The only known solution is to exploit the difference in polari-
zation between blue sky and white clouds. Light from most por-
tions of the blue sky is partially linearly polarized, as explained
in a preceding section, and light from clouds is unpolarized.
Therefore a neutral-color, linear polarizer mounted at the opti-
mum azimuth in front of the lens will absorb a large fraction
(e.g., 80 percent) of the sky light while transmitting a large frac-
76
Popular Applications of Polarized Light
tion (nearly half) of the light from the clouds; thus the contrast
is increased by a factor of two or three. The factor is less if the
air is hazy, and more if the air is extremely clear (as in Arizona)
and if the camera is aimed about 90° from the direction of the
sun.
The usual way of choosing the azimuth of the polarizer is
crude, but perhaps adequate. The photographer holds the polar-
izer in front of his eye, finds by trial and error which azimuth
maximizes the contrast of the clouds in question, and then at-
tempts to mount the polarizer on the camera without changing
the azimuth of the polarizer. One type of polarizing filter for
cameras is equipped with a small "satellite" polarizer mounted
at the end of a short arm and aligned permanently with the main
polarizer. The photographer installs the main polarizer in front
of the lens, looks through the small polarizer and turns the arm
to whatever azimuth maximizes the contrast. Both polarizers
then have this optimum orientation. The satisfactoriness of the
azimuth can be checked visually at any time. Instead of using
these empirical methods, a scientifically minded photographer
can proceed by dead reckoning, i.e., by following this well-known
rule: Mount the polarizer so that its transmission axis lies in the
plane determined by camera, sun, and object photographed. (So
oriented, the polarizer performs a valuable additional service: it
eliminates most of the specularly reflected light from trees, roads,
etc., and enhances the softness and depth of color of the scene.)
When a photographer standing on a sidewalk tries to photo-
graph objects situated behind a store window, the reflection of
the street scene from the window may threaten to spoil the
photograph. An excellent solution is to place the camera off to
one side so that the window is seen obliquely at about the polar-
izing angle, and mount a linear polarizer in front of the lens;
the polarizer is turned so that its transmission axis is horizontal,
and the polarized light reflected from the window is absorbed.
The authors have a friend who has applied this same principle
to a pair of special spectacles he wears while touring the country
by railroad. The lenses consist of polarizers oriented with the
transmission axis horizontal, as indicated in Fig. 10-2b; thus
when he gazes out of the train window in oblique forward direc-
77
tion, the reflected images of passengers and newspapers are wiped
out, and the scenery appears in its pristine glory.
USE OF CIRCULAR POLARIZERS IN ELIMINATING
PERPENDICULARLY REFLECTED LIGHT
Eliminating perpendicularly reflected lights is a different prob-
lem from that of eliminating obliquely reflected light. The proc-
ess of oblique reflection at Brewster's angle causes the reflected
beam to be linearly polarized, and accordingly a linear polarizer
can eliminate the reflected beam entirely. But the process of
normal reflection, i.e., with incident and reflected beams perpen-
dicular to the smooth glossy surface in question, produces no
polarization at all. How, then, can the specularly reflected light
be eliminated while light originating behind the surface is trans-
mitted freely?
The question is an important one to radar operators scanning
the cathode-ray-oscilloscope screens on which dim greenish spots
representing airborne objects appear. The screen proper is situ-
ated in a large evacuated tube, and the greenish light emerges
through a curved glass window at the front end of the tube.
(Sometimes the window is flat; sometimes a safety plate of glass
or plastic is mounted close in front of it.) Often the operator
has difficulty in seeing the greenish spots, not only because they
are faint, but also because they may be masked by various ex-
traneous images reflected by the front surface of the window, e.g.,
reflections of room lights and of people, clothing, papers, etc.,
situated near the operator. Extinguishing the room lights would
eliminate these reflections, but would make it impossible for
the operator to read instructions or make notes. What he needs
is some kind of filter that will transmit the light originating be-
hind the window and absorb the light reflected approximately
perpendicularly from it.
This need is filled by the circular polarizer. Such a device, if
mounted close in front of the window, will transmit nearly half
of the light that originates behind the window, yet will eliminate
about 99 percent of the room light that is reflected perpen-
dicularly from it. The circular polarizer acts on the room light
twice: it circularly polarizes room light that is approaching the
78
Popular Applications of Polarized Light
window, then absorbs the reflected component. The logic behind
this requires explanation. Two key facts must be kept in mind:
(1) A beam that is reflected perpendicularly and specularly by
a smooth glossy surface has the same degree of polarization as
the incident beam, since the reflection process does not intro-
duce randomness of any kind.
(2) The reflection process reverses the handedness of polariza-
tion, because handedness is defined with respect to the propaga-
tion direction and the reflection process reverses the propagation
direction.
If the polarizer is of right-circular type, as in the arrangement
shown in Fig. 10-3, room light that passes through and ap-
UNPOLARIZED
LIGHT
RETARDING
LAYER
SPECULARLY
REFLECTING
GLASS PLATE/7
POLARIZER
FIG. 10-3 Use of a circular polarizer in absorbing light reflected by a
surface approximately perpendicular to the incident beam. Note that
the reflection process reverses the handedness of circular polarization.
proaches the window is right-circularly polarized; the reflected
light is /e/f-circularly polarized and hence is totally absorbed by
the polarizer. In effect, the circular polarizer "codes" the light,
the window reverses the coding, and the polarizer then annihi-
lates the reverse-coded beam. If both faces of the window are
ideally flat and smooth, if the light is incident exactly along
the normal, and if the polarizer is truly of circular type, the
reflected light is totally absorbed. Usually the conditions are
less ideal: the rear surface of the window usually serves as sup-
port for the luminescent screen and has a matte appearance; the
window is usually curved and much of the troublesome room
light incident on the window makes an angle of 10° or 20° or
more with the normal; and the polarizer, although circular with
respect to some wavelengths, is elliptical with respect to others.
Nevertheless, the improvement provided by the polarizer is large,
and the amount of faint detail that the operator can see on the
screen is greatly increased.
One precaution must be mentioned: reflections from the polar-
izer itself must be avoided. This is usually accomplished by tilt-
ing the polarizer forward so that the only reflected images the
observer sees are images of a dark-colored floor or other dark
objects.
Television sets, also, have been equipped successfully with cir-
cular polarizers. If the set is used in a brightly lit room, or is used
outdoors, the circular polarizer performs a valuable service in
trapping the specularly reflected glare and thus increasing the
picture-vs-glare ratio by a factor of the order of 10.
VARIABLE-DENSITY FILTER
A pair of linear polarizers arranged in series is an almost ideal
device for controlling the transmitted intensity of light. Rotating
one polarizer through an angle 6 with respect to the other causes
the intensity of the transmitted light to vary approximately as
cos2 6. Because the transmittance is easily varied and easily calcu-
lated, the pair of polarizers has found much favor in the eyes of
designers of spectrophotometers and other devices for controlling
and measuring light intensity.
Specially designed sunglasses employing pairs of linear polar-
izers in place of lenses have been used successfully by aviators
and others. One polarizer of each pair can be rotated through
an angle as large as 90°, and a linkage connecting the two pairs
insures that the attenuation is the same for both eyes. By moving
one small lever, the wearer can vary the transmittance through-
out a range of about 10,000 to 1.
Controllable pairs of very-large-diameter polarizers have been
80
Popular Applications of Polarized Light
used as windows of railroad cars and ocean liners. A person
sitting near such a window turns a small knob to rotate one
polarizer with respect to the other and thus reduce the intensity
of the transmitted light to any extent desired.
One of the authors has experimented with a variable-density
filter employing three linear polarizers in series, in order that a
transmittance range of 108 to 1 could be achieved. The device
worked well and, as expected, obeyed a cosine-fourth, rather than
a cosine-square law.
THREE-DIMENSIONAL PHOTOGRAPHY AND THE USE
OF POLARIZERS FOR CODING
Millions of polarizers found their way into the motion picture
theaters of North America in 1952 and 1953 when stereoscopic
(three-dimensional, or 3-D) movies achieved brief prominence.
Each spectator wore a pair of polarizing spectacles called view-
ers, and polarizers were mounted in front of the projectors.
A photographer who enjoys looking at 3-D still pictures in his
living room needs no polarizers. Usually he employs a small view-
ing box containing a light source and two lenses, one for each
eye; a black partition, or septum, divides the box into right and
left halves. The picture, consisting of two small transparencies
mounted about two inches apart in a side-by-side arrangement on
a cardboard frame, is inserted in the box so that the right-eye
transparency lines up with the right lens and the left-eye trans-
parency lines up with the left lens. (The two transparencies are,
of course, slightly different because they were taken by cameras
situated about two or three inches apart; the spacing used ap-
proximates the spacing of the two eyes.) The side-by-side arrange-
ment of the two transparencies and the presence of the septum
insure that the observer's right eye sees only the right trans-
parency and the left eye sees only the left transparency. No
cross-communication, or "cross-talk," can occur. Consequently
the observer enjoys an impressively realistic stereoscopic illusion.
When 3-D motion-picture films are projected in a theater,
many complications arise. Separate projectors must be provided
for the right-eye and left-eye movie films, and the two projectors
must be synchronized within about 0.01 second. Since there is
81
just one large screen and this is to be viewed by hundreds of
spectators, there can be no septum. Indeed, no practical geometri-
cal method of preventing cross-talk is known.
Before the advent of mass-produced polarizers in the 1930's, an
analglyph system of preventing cross-talk was invented. It applied
wavelength coding to the two projected beams. The right-eye
picture was projected through a long-wavelength (red) filter, and
the left-eye picture was projected through a shorter-wavelength
(green) filter. The spectator's viewers contained right and left
lenses of red and green plastic, respectively, and accordingly each
lens transmitted light from the appropriate projector and ab-
sorbed light from the other. Thus each eye received just the
light intended for it. The system succeeded as a short-term
novelty: stereoscopic illusions were created. But the system had
two major defects: chromatic "retinal rivalry" between the two
eyes, and incompatibility with the showing of colored motion
pictures.
In the 1930's the problem was solved with £clat by a polariza-
tion-coding system, demonstrated with great impact at the New
York World's Fair of 1939 and improved in later years. As indi-
cated in Fig. 10-4, a linear polarizer oriented with its transmis-
sion axis at —45° is placed in front of the projector used for the
right-eye pictures, and a polarizer at +45° is placed in front of
the projector used for the left-eye pictures. Thus the two beams
striking the movie screen are orthogonally coded. The lenses
of the spectator's viewers consist of correspondingly oriented
linear polarizers, and so each eye receives only light that origi-
nates in the appropriate projector. Superb stereoscopic illusions
result. Since the polarizers perform well at all wavelengths in the
visual range, color movies can be presented as easily and faith-
fully as can black-and-white movies.
The polarizers placed in front of the projectors consist, ordi-
narily, of K-sheet; as explained in Chapter 3, K-sheet is highly
resistant to heat, and any polarizing filter placed close in front
of a powerful projector is bound to heat up considerably since it
necessarily absorbs about half the light. The lenses of the 3-D
viewers are usually of HN-38 sheet; it has high major trans-
mittance kx and small minor transmittance k2, and it is inex-
82
Popular Applications of Polarized Light
FIG. 10-4 Arrangement for projecting polarization-coded stereoscopic
motion-picture films by means of two side-by-side projectors. Films FR
and FL containing the "right-eye pictures" and "left-eye pictures" are
mounted in the right and left projectors, which are equipped with
linear polarizers PR and PL oriented at —45° and +45* respectively.
The viewer contains correspondingly oriented polarizers, and accord-
ingly each eye sees only the images intended for it.
pensive. The viewers are cheap enough (about 10^ each) that
they can be discarded after a single use.
The polarization-coding scheme has one limitation: if the
spectator tilts his head to one side, the polarizers in his viewers
no longer line up accurately with the respective polarizers on the
projectors. Thus cross-talk occurs: the right eye sees faintly the
image meant for the left eye, and vice versa: each eye sees a
faint ghost image in addition to the main image. The spectator
does not enjoy this. The difficulty could be avoided if the linear
polarizers were replaced by high-quality, achromatic circular
polarizers, but unfortunately no method is known for producing
achromatic circular polarizers economically.
The effectiveness of any polarization-coding projection system
is destroyed if the screen depolarizes the light appreciably.
Screens that have a smooth aluminum coating usually conserve
83
polarization to the extent of about 99 percent, but those having
a matte white surface or a rough metallic coating produce much
depolarization and hence much cross-talk between the two
images. Many of the screens used in the innocent days of 1952
and 1953 were of the wrong type, and the resulting ghost images
were a major annoyance. For that reason, and because of fre-
quent lack of care in maintaining synchronism between the two
projectors, movie-goers socn turned back to conventional 2-D
pictures. Some nostalgia remains, however. Persons who were
lucky enough to see a full-color, 3-D movie showing attractive
actors filmed against a background of gorgeous scenery look for-
ward to the time when well-made, well-presented 3-D movies,
with their almost miraculous realism and intimacy, will animate
the theaters once again.
THE VECTOGRAPH
The type of three-dimensional photography discussed in the
preceding section is parallel-projected 3-D photography. The
two motion-picture films are situated side-by-side, and two pro-
jectors are operated in parallel. During the late 1930's a radically
new approach, called vcctography, was developed by E. H. Land,
J. Mahler, and others. In this system, the two films are arranged
in series, bonded together. Because of the permanent series ar-
rangement, many problems disappear. Only one projector is
needed, and perfect synchronism is "guaranteed at the factory."
Each pair of pictures (each vectograph) is projected as a single
unit, in the same projector aperture and at the same time, and
onto the same area of the same screen. If the film breaks, it can
be spliced with no concern as to preservation of synchronism.
The method can succeed only if means are provided for pre-
serving the identity of the two coincident projected beams.
Again, polarization-coding is the answer. However, because the
two images are bonded together in series, the coding must occur
within the images themselves. In the system used by Land and
Mahler each image consists of varying quantities of linearly
dichroic molecules aligned in a common direction, and the direc-
tions employed in the two images are mutually at right angles.
Dark areas in any one image contain a high concentration of
84
Popular Applications of Polarized Light
dichroic molecules; light areas contain little or no dichroic
material; but irrespective of concentration, the alignment direc-
tion is always the same. For the other image, the alignment direc-
tion is always orthogonal to the first. It is to be noted that the
images contain no silver and no other isotropic absorber. Only
aligned absorbers having high dichroic ratio are used.
A communications engineer would describe the vectograph by
saying that it provides two distinct channels. Each is assigned to
one image. Each is independent of the other. Since the vecto-
graph images themselves perform the polarization coding, no
polarizer is used in front of the projector; indeed, the interpo-
sition of such a filter would play havoc with the system. As be-
fore, the screen must preserve the polarization and the spectator's
viewers must perform the appropriate decoding, or discrimi-
nating, act. Excellent stereoscopic effects are achieved. However,
the production of vectograph film is a costly undertaking involv-
ing very specialized equipment, and constant attention is needed
to maintain high enough dichroic ratio so that the channels are
truly independent and ghost images are avoided.
Vectograph pictures of the "still" type are easier and cheaper
to make than vectograph movies. Stereo pairs of aerial photo-
graphs of mountainous country, if presented in vectograph form,
give a navigator (wearing an appropriate viewer) a very realistic
impression of the terrain, and a map maker can prepare an accu-
rate contour map from the vectograph with ease.
POLARIZING HEADLIGHTS
It is ironic that the main goal of Land and others in develop-
ing high-quality, large-area, low-cost polarizers has never been
achieved. The polarizers are used with great success in dozens
of applications, but not the application that was uppermost in
the minds of the inventors.
Their goal was to eliminate glare from automobile headlights.
In an era when dual-lane highways, circumferential bypasses,
and other safety engineering advances were virtually unknown
and the aim and focus of automobile headlights were highly
erratic, the glare that confronted motorists at night was almost
85
unbearable, and was an important cause of accidents. As early
as 1920 several illumination engineers recognized that the glare
could be eliminated by means of polarizers — if large-area polar-
izers could somehow be produced. If every headlight lens were
covered by a linear polarizer oriented with the transmission
axis horizontal and every windshield were covered with a linear
polarizer oriented with its axis vertical, no direct light from the
headlights of Car A could pass through the windshield of on-
coming Car B. Drivers in both cars could see road-markings,
pedestrians, and so forth, but neither would experience any
glare from the other's headlights. Moreover, it would be per-
missible for each driver to use his high beam continuously, and
accordingly his ability to see pedestrians, etc., would be greater
than before, despite the fact that each polarizer would transmit
only about half of the light incident on it.
It was soon recognized that the analyzing polarizer should not
be made a permanent part of the windshield, but should be in-
corporated in a small visor situated just in front of the driver's
eyes. During the day, when headlights were not in use, the visor
could be swung out of the way. It was also recognized that care
should be taken to make sure the headlight polarizers had suffi-
cient light-leak, i.e., sufficiently large k2 value, that the head-
lights of oncoming cars would not disappear entirely!
Land and his colleagues moved rapidly. They invented a
whole series of polarizers, each superior to its predecessor. The
first successful type, J-sheet, employed aligned, microscopic crys-
tals of the dichroic mineral herapathite; the method of manu-
facture is described in Chapter 3. Then came H-sheet, which
was better in nearly every respect and in addition was easier to
make. Finally, K-sheet appeared; it had most of the superb
qualities of the earlier materials and the added virtue of being
unaffected by fairly high temperature, such as 215°F. To per-
sons seeking polarizers for use in headlights, K-sheet appeared
to be the pot of gold at the end of a polarized rainbow.
Concurrently, several better ways of orienting the polarizers
were proposed. One attractive scheme was to orient the head-
light polarizers and the visor polarizer at the identical azimuth,
namely —45°, as indicated in Fig. 10-5. Then, even a polariza-
tion-conserving object in the path of the headlights would appear
86
Popular Applications of Polarized Light
FIG. 10-5 Automobile equipped with headlight polarizers and a visor
polarizer oriented at —45°. When two such cars approach one another,
each driver is protected from the glare from the headlights of the other.
to the driver to be brightly illuminated. (This would not be the
case if his visor polarizer were crossed with his headlight polar-
izers.) The —45° system disposed of the headlight glare problem
adequately: if two cars A and B both equipped in this manner
approached one another at night, each driver's visor would be
crossed with the other car's headlight polarizers, and neither
driver would experience any glare.
Using the Mueller calculus, Billings and Land compared a
wide variety of polarizer orientation schemes, and found several
to be particularly attractive. Perhaps the best system was one
called " — 55°, —35°." The transmission axes of the headlight
polarizers and visor polarizer are at 55° and 35° from the vertical,
respectively, an arrangement that minimizes complications stem-
ming from the obliquity of the portion of the windshield situated
just in front of the driver.
Despite the successes on all technical fronts, the project bogged
down. To this day no one knows just why. Probably many little
reasons were responsible. Among these were the following:
(1) The polarizers absorbed slightly more than half of the
light incident on them, and accordingly the automobile manu-
facturers felt that they would have to increase the power of the
lamps themselves and perhaps use larger generators and batteries
also.
(2) Some windshields were moderately birefringent; therefore
87
they would act like retarders, alter the polarization form of the
incident light, and allow some glare to leak through.
(3) Nearly every year the automobile manufacturers increased
the backward tilt of the windshields; such tilt tends to alter the
polarization form of light having an oblique vibration direction,
and hence leads to glare-leak.
(4) Passengers, as well as drivers, would require visors, since
passengers also dislike glare.
(5) Pedestrians might find that the glare was worse than ever,
unless they too employed polarizing visors or spectacles.
(6) The system would succeed only if adopted by all car manu-
facturers, and therefore no one manufacturer would gain any
promotional advantage from it.
(7) The first few drivers to put the system to use would get
little benefit from it for at least a year or two, i.e., until millions
of other cars were similarly equipped.
(8) It was difficult to decide when and how to force the owners
of old cars to install the necessary polarizers on their cars.
(9) The patents on the only fully satisfactory polarizers were
held by a single company.
(10) To introduce the system would require formal, coordi-
nated action by all States.
(11) Improvements in headlight design and aiming, the in-
creasing numbers of dual-lane highways, and the brighter street
lamps used in cities and suburbs led some people to believe that
the need for a polarization-type of glare control was no longer
acute.
However, persons who have actually experienced the polariza-
tion method of glare removal are convinced that the drawbacks
are trivial compared to the benefits.
Perhaps some day the system will be tried out on a pilot scale
in a small, isolated community, where all the cars could be
equipped with polarizers in a few weeks. Perhaps an island of
moderate size would make a good test ground. If the system is
found to be highly successful there, it will presumably spread
throughout every country that teems with automobiles.
88
An explanation of how the eye works, by the
biologist who won a Nobel Prize for contributions to this field.
Eye and Camera
George Wald
A Scientific American article, 1950
OF all the instruments made by man,
none resembles a part of his body
more than a camera does the eye.
Yet this is not by design. A camera is
no more a copy of an eye than the wing
of a bird is a copy of that of an insect.
Each is the product of an independent
evolution; and if this has brought the
camera and the eye together, it is not
because one has mimicked the other, but
because both have had to meet the same
problems, and frequently have done so
in much the same way. This is the type
of phenomenon that biologists call con-
vergent evolution, yet peculiar in that
the one evolution is organic, the other
technological.
Over the centuries much has been
learned about vision from the camera,
but little about photography from the
eye. The camera made its first appear-
ance not as an instrument for making
pictures but as the camera obscura or
dark chamber, a device that attempted
no more than to project an inverted
image upon a screen. Long after the
optics of the camera obscura was well
understood, the workings of the eye re-
mained mysterious.
In part this was because men found
it difficult to think in simple terms about
the eye. It is possible for contempt to
breed familiarity, but awe does not help
one to understand anything. Men have
often approached light and the eye in a
spirit close to awe, probably because
they were always aware that vision pro-
vides their closest link with the external
world. Stubborn misconceptions held
back their understanding of the eye for
many centuries. Two notions were par-
ticularly troublesome. One was that ra-
diation shines out of the eye; the other,
that an inverted image on the retina is
somehow incompatible with seeing right
side up.
I am sure that many people are still
not clear on either matter. I note, for
example, that the X-ray vision of the
comic-strip hero Superman, while re-
garded with skepticism by many adults,
is not rejected on the ground that there
are no X-rays about us with which to
see. Clearly Superman's eyes supply the
X-rays, and by directing them here and
there he not only can see through opaque
objects, but can on occasion shatter a
brick wall or melt gold. As for the in-
verted image on the retina, most people
who learn of it concede that it presents
a problem, but comfort themselves with
the thought that the brain somehow
compensates for it. But of course there
is no problem, and hence no compensa-
tion. We learn early in infancy to asso-
ciate certain spatial relations in the
outside world with certain patterns of
nervous activity stimulated through the
eyes. The spatial arrangements of the
nervous activity itself are altogether
irrelevant.
It was not until the 17th century that
the gross optics of image formation in
the eye was clearly expressed. This was
accomplished by Johannes Kepler in
1611, and again by Rene Descartes in
1664. By the end of the century the first
treatise on optics in English, written by
William Molyneux of Dublin, contained
several clear and simple diagrams com-
paring the projection of a real inverted
image in a "pinhole" camera, in a cam-
era obscura equipped with a lens and
in an eye.
Today every schoolboy knows that the
eye is like a camera. In both instruments
a lens projects an inverted image of the
surroundings upon a light-sensitive sur-
face: the film in the camera and the
retina in the eye. In both the opening
of the lens is regulated by an iris. In
both the inside of the chamber is lined
with a coating of black material which
absorbs stray light that would otherwise
be reflected back and forth and obscure
the image. Almost every schoolboy also
knows a difference between the camera
and the eye. A camera is focused by mov-
ing the lens toward or away from the
film; in the eye the distance between the
lens and the retina is fixed, and focusing
is accomplished by changing the thick-
ness of the lens.
The usual fate of such comparisons is
that on closer examination they are ex-
posed as trivial. In this case, however,
just the opposite has occurred. The more
we have come to know about the mech-
anism of vision, the more pointed and
fruitful has become its comparison with
photography. By now it is clear that the
relationship between the eye and the
camera goes far beyond simple optics,
and has come to involve much of the
89
CONJUNCTIVA
FOVEA
CILIARY
MUSCLE
NERVE
SCLERA CHOROID RETINA
''>"/ /ti\\\\\\\\\
OPTICAL SIMILARITIES of eye and camera are ap-
|iurent in their cross sections. Both utilize ;i lens to
funis an inverted image on a light-sensitive surface.
Botli posses* .in iris to adjust to various intensities of
liplit. The single lens of the eye, however, cannot bring
li^lit of all colors to a focus at the same point. The
compound lens of the camera is better corrected for
color because it is composed of two kinds of glass.
90
WMtf
FORMATION OF AN* IMAGE on the retina of the human eye was diagrammed
hy Rene Descartes in 1664. This diagram is from Descartes' Dioptrics.
Eye and Camera
essential physics and chemistry of both
devices.
Bright and Dim Light
A photographer making an exposure
in dim light opens the iris of his camera.
The pupil of the eye also opens in dim
light, to an extent governed by the activ-
ity of the retina. Both adjustments have
the obvious effect of admitting more
light through the lens. This is accom-
plished at some cost to the quality of
the image, for the open lens usually de-
fines the image less sharply, and has less
depth of focus.
When further pressed for light, the
photographer changes to a more sensi-
tive film. This ordinarily involves a fur-
ther loss in the sharpness of the picture.
With any single type of emulsion the
more sensitive film is coarser in grain,
and thus the image cast upon it is re-
solved less accurately.
The retina of the eye is grainy just as
is photographic film. In film the grain is
composed of crystals of silver bromide
embedded in gelatin. In the retina it is
made up of the receptor cells, lying side
by side to form a mosaic of light-sensitive
elements.
There are two kinds of receptors in the
retinas of man and most vertebrates:
rods and cones. Each is composed of an
inner segment much like an ordinarv
nerve cell, and a rod- or cone-shaped
outer segment, the special portion of the
cell that is sensitive to light. The cones
are the organs of vision in bright light,
and also of color vision. The rods provide
a special apparatus for vision in dim
light, and their excitation yields only
neutral gray sensations. This is why at
night all cats are gray.
The change from cone to rod vision,
like that from slow to fast film, involves
a change from a fine- to a coarse-grained
mosaic. It is not that the cones are
smaller than the rods, but that the cones
act individually while the rods act in
large clumps. Each cone is usually con-
nected with the brain by a single fiber
of the optic nerve. In contrast large
clusters of rods are connected by single
optic nerve fibers. The capacity of rods
for image vision is correspondingly
coarse. It is not only true that at night
all cats are gray, but it is difficult to be
sure that they are cats.
Vision in very dim light, such as star-
light or most moonlight, involves only
the rods. The relatively insensitive cones
are not stimulated at all. At moderately
low intensities of light, about 1,000 times
greater than the lowest intensity to
which the eye responds, the cones begin
to function. Their entrance is marked bv
dilute sensations of color. Over an inter-
mediate range of intensities rods and
cones function together, but as the
brightness increases, the cones come to
dominate vision. We do not know that
91
SSfig&S
GRAIN of the photographic emulsion, magnified 2,500
times, is made up of silver-bromide crystals in gelatin.
"GRAIN" of the human retina is made up of cones and
rods (dots at far right). Semicircle indicates fovea.
the rods actually stop functioning at
even the highest intensities, but in bright
light their relative contribution to vision
falls to so low a level as to be almost
negligible.
To this general transfer of vision from
rods to cones certain cold-blooded ani-
mals add a special anatomical device.
The light-sensitive outer segments of the
rods and cones are carried at the ends of
fine stalks called myoids, which can
shorten and lengthen. In dim light the
rod myoids contract while the cone my-
oids relax. The entire field of rods is thus
pulled forward toward the light, while
the cones are pushed into the back-
ground. In bright light the reverse oc-
curs: the cones are pulled forward and
the rods pushed back. One could scarce-
ly imagine a closer approach to the
change from fast to slow film in a camera.
The rods and cones share with the
grains of the photographic plate another
deeply significant property. It has long
been known that in a film exposed to
light each grain of silver bromide given
enough developer blackens either com-
pletely or not at all, and that a grain is
made susceptible to development by the
absorption of one or at most a few quanta
of light. It appears to be equally true
that a cone or rod is excited by light to
yield either its maximal response or none
at all. This is certainly true of the nerve
fibers to which the rods and cones are
connected, and we now know that to
produce this effect in a rod— and possibly
also in a cone— only one quantum of light
need be absorbed.
It is a basic tenet of photochemistry
that one quantum of light is absorbed by,
and in general can activate, only one
molecule or atom. We must attempt to
understand how such a small beginning
can bring about such a large result as the
development of a photographic grain
or the discharge of a retinal receptor.
In the photographic process the answer
to this question seems to be that the ab-
sorption of a quantum of light causes the
oxidation of a silver ion to an atom of
metallic silver, which then serves as a
catalytic center for the development of
the entire grain. It is possible that a
similar mechanism operates in a rod or a
cone. The absorption of a quantum of
light by a light-sensitive molecule in
either structure might convert it into a
biological catalyst, or an enzyme, which
could then promote the further reactions
that discharge the receptor cell. One
wonders whether such a mechanism
could possibly be rapid enough. A rod
or a cone responds to light within a
small fraction of a second; the mecha-
nism would therefore have to complete
its work within this small interval.
One of the strangest characteristics of
the eye in dim light follows from some
of these various phenomena. In focusing
the eye is guided by its evaluation of the
sharpness of the image on the retina. As
the image deteriorates with the opening
of the pupil in dim light, and as the
retinal capacity to resolve the image falls
with the shift from cones to rods, the
ability to focus declines also. In very
dim light the eye virtually ceases to ad-
just its focus at all. It has come to resem-
ble a very cheap camera, a fixed-focus
instrument.
In all that concerns its function, there-
fore, the eye is one device in bright light
and another in dim. At low intensities all
its resources are concentrated upon sen-
sitivity, at whatever sacrifice of form; it
is predominantly an instrument for see-
ing light, not pattern. In bright light all
this changes. By narrowing the pupil,
shifting from rods to cones, and other
stratagems still to be described, the eye
sacrifices light in order to achieve the
utmost in pattern vision.
Images
In the course of evolution animals
have used almost everv known device
for forming or evaluating an image.
There is one notable exception: no ani-
mal has yet developed an eye based
upon the use of a concave mirror. An
eye made like a pinhole camera, how-
ever, is found in Nautilus, a cephalopod
mollusk related to the octopus and squid.
The compound eye of insects and crabs
forms an image which is an upright
patchwork of responses of individual
"eyes" or ommatidia, each of which
records only a spot of light or shade.
The eye of the tiny arthropod Copilia
possesses a large and beautiful lens but
only one light receptor attached to a thin
strand of muscle. It is said that the mus-
cle moves the receptor rapidly back and
forth in the focal plane of the lens, scan-
ning the image in much the same way as
it is scanned by the light-sensitive tube
of a television camera.
Each of these eyes, like the lens eye
of vertebrates, represents some close
compromise of advantages and limita-
tions. The pinhole eye is in focus at all
distances, yet to form clear images it
must use a small hole admitting very
little light. The compound eye works
well at distances of a few millimeters,
yet it is relatively coarse in pattern reso-
lution. The vertebrate eye is a long-
range, high-acuitv instrument useless in
the short distances at which the insect
eye resolves the greatest detail.
These properties of the vertebrate eye
are of course shared by the camera. The
use of a lens to project an image, how-
ever, has created for both devices a spe-
cial group of problems. All simple lenses
are subject to serious errors in image
formation: the lens aberrations.
Spherical aberration is found in all
lenses bounded by spherical surfaces.
The marginal portions of the lens bring
rays of light to a shorter focus than the
central region. The image of a point in
space is therefore not a point, but a little
"blur circle." The cost of a camera is
largely determined by the extent to
92
CONES of the catfish Ameiurus are
pulled toward the surface of the ret-
ina (top) in bright light. The rods
remain in a layer below the surface.
RODS advance and cones retreat in
dim light. This retinal feature is not
possessed by mammals. It is peculiar
to some of the cold-blooded animals.
which this aberration is corrected by
modifying the lens.
The human eye is astonishingly well
corrected— often slightly overcorrected—
for spherical aberration. This is accom-
plished in two ways. The cornea, which
is the principal refracting surface of the
eye, has a flatter curvature at its margin
than at its center. This compensates in
part for the tendency of a spherical sur-
face to refract light more strongly at its
margin. More important still, the lens is
denser and hence refracts light more
strongly at its core than in its outer
layers.
A second major lens error, however,
remains almost uncorrected in the hu-
man eye. This is chromatic aberration,
or color error. All single lenses made of
one material refract rays of short wave-
length more strongly than those of longer
wavelength, and so bring blue light to a
shorter focus than red. The result is that
the image of a point of white light is not
a white point, but a blur circle fringed
with color. Since this seriously disturbs
the image, even the lenses of inexpensive
cameras are corrected for chromatic
aberration.
It has been known since the time of
Isaac Newton, however, that the human
eye has a large chromatic aberration. Its
lens system seems to be entirely uncor-
rected for this defect. Indeed, living or-
ganisms are probably unable to manu-
facture two transparent materials of such
widely different refraction and disper-
sion as the crown and flint glasses from
which color-corrected lenses are con-
structed.
The large color error of the human eye
could make serious difficulties for image
vision. Actually the error is moderate
between the red end of the spectrum and
the blue-green, but it increases rapidly
at shorter wavelengths: the blue, violet
and ultraviolet. These latter parts of the
spectrum present the most serious prob-
lem. It is a problem for both the eye and
the camera, but one for which the eye
must find a special solution.
The first device that opposes the color
error of the human eye is the yellow lens.
The human lens is not only a lens but a
color filter. It passes what we ordinarily
consider to be the visible spectrum, but
sharply cuts off the far edge of the violet,
in the region of wavelength 400 milli-
microns. It is this action of the lens, and
not any intrinsic lack of sensitivity of
the rods and cones, that keeps us from
seeing in the near ultraviolet. Indeed,
persons who have lost their lenses in the
operation for cataract and have had
them replaced by clear glass lenses, have
excellent vision in the ultraviolet. They
are able to read an optician's chart from
top to bottom in ultraviolet light which
leaves ordinary people in complete
darkness.
The lens therefore solves the problem
of the near ultraviolet, the region of the
Eye and Camera
spectrum in which the color error is
greatest, simply by eliminating the re-
gion from human vision. This boon is
distributed over one's lifetime, for the
lens becomes a deeper yellow and makes
more of the ordinary violet and blue in-
visible as one grows older. I have heard
it said that for this reason aging artists
tend to use less blue and violet in their
paintings.
The lens filters out the ultraviolet for
the eye as a whole. The remaining de-
vices which counteract chromatic aber-
ration are concentrated upon vision in
bright light, upon cone vision. This is
good economy, for the rods provide such
a coarse-grained receptive surface that
they would be unable in any case to
evaluate a sharp image on the retina.
As one goes from dim to bright light,
from rod to cone vision, the sensitivity
of the eye shifts toward the red end of
the spectrum. This phenomenon was de-
scribed in 1825 by the Czech physiolo-
gist Johannes Purkinje. He had noticed
that with the first light of dawn blue ob-
jects tend to look relatively bright com-
pared with red, but that they come to
look relatively dim as the morning ad-
vances. The basis of this change is a
large difference in spectral sensitivity
between rods and cones. Rods have their
maximal sensitivity in the blue-green at
about 500 millimicrons; the entire spec-
tral sensitivity of the cones is transposed
toward the red, the maximum lying in
the yellow-green at about 562 millimi-
crons. The point of this difference for
our present argument is that as one goes
from dim light, in which pattern vision
is poor in any case, to bright light, in
which it becomes acute, the sensitivity
of the eye moves away from the region
of the spectrum in which the chromatic
aberration is large toward the part of the
spectrum in which it is least.
The color correction of the eye is com-
pleted by a third dispensation. Toward
the center of the human retina there is a
small, shallow depression called the fo-
vea, which contains only cones. While
the retina as a whole sweeps through a
visual angle of some 240 degrees, the
fovea subtends an angle of only about
1.7 degrees. The fovea is considerably
smaller than the head of a pin, yet with
this tiny patch of retina the eye accom-
plishes all its most detailed vision.
The fovea also includes the fixation
point of the eye. To look directly at
something is to turn one's eye so that
its image falls upon the fovea. Beyond
the boundary of the fovea rods appear,
and they become more and more nu-
merous as the distance from the fovea
increases. The apparatus for vision in
bright light is thus concentrated toward
the center of the retina, that for dim
light toward its periphery. In very dim
light, too dim to excite the cones, the
fovea is blind. One can see objects then
only by looking at them slightly askance
93
to catch their images on areas rich in
rods.
In man, apes and monkeys, alone of
all known mammals, the fovea and the
region of retina just around it is colored
yellow. This area is called the yellow
patch, or macula lutca. Its pigmentation
lies as a yellow screen over the light re-
ceptors of the central retina, subtending
a visual angle some five to 10 degrees in
diameter.
Several years ago in our laboratory at
Harvard University we measured the
color transmission of this pigment in the
living human eye by comparing the spec-
tral sensitivities of cones in the yellow
patch with those in a colorless peripheral
area. The yellow pigment was also ex-
tracted from a small number of human
maculae, and was found to be xaniho-
phyll, a carotenoid pigment that occurs
also in all green leaves. This pigment in
the yellow patch takes up the absorption
of light in the violet and blue regions of
the spectrum just where absorption by
the lens falls to very low values. In this
way the yellow patch removes for the
central retina the remaining regions of
the spectrum for which the color error is
high.
So the human eye, unable to correct
its color error otherwise, throws away
those portions of the spectrum that
would make the most trouble. The yel-
low lens removes the near ultraviolet for
the eye as a whole, the macular pigment
eliminates most of the violet and blue
for the central retina, and the shift from
rods to cones displaces vision in bright
light bodily toward the red. By these
three devices the apparatus of most
acute vision avoids the entire range of
the spectrum in which the chromatic
aberration is large.
Photography with Living Eyes
In 1876 Franz Boll of the University
of Rome discovered in the rods of the
frog retina a brilliant red pigment. This
bleached in the light and was resynthe-
sized in the dark, and so fulfilled the
elementary requirements of a visual pig-
ment. He called this substance visual
red; later it was renamed visual purple
or rhodopsin. This pigment marks the
point of attack by light on the rods: the
absorption of light by rhodopsin initiates
the train of reactions that end in- rod
vision.
Boll had scarcely announced his dis-
covery when Willy Kiihne, professor of
physiology at Heidelberg, took up the
.study of rhodopsin, and in one extraor-
dinary year learned almost everything
about it that was known until recently.
In his first paper on retinal chemistry
Kiihne said: "Bound together with the
pigment epithelium, the retina behaves
not merely like a photographic plate, but
like an entire photographic workshop, in
which the workman continually renews
PINHOLE-CAMERA EYE is found in Nautilus, the spiral-shelled mollusk
which is related to the octopus and the squid. This eye has the advantage
of being in focus at all distances from the object that is viewed. It has the
serious disadvantage, however, of admitting very little light to the retina.
COMPOUND EYE is found in insects. Each element contributes only a small
patch of light or shade to make up the whole mosaic image. This double
compound eye is found in the mayfly Chloeon. The segment at the top
provides detailed vision; the segment at the right, coarse, wide-angled vision.
SCANNING EYE is found in the arthropod Copilia. It possesses a large lens
I right I but only one receptor element (left). Attached to the receptor
are the optic nerve and a strand of muscle. The latter is reported to move
the receptor back and forth so that it scans the image formed by the len9.
94
Eye and Camera
SPHERICAL ABERRATION occurs when light is refracted hy a lens with
spherical surfaces. The light which passes through the edge of the lens is
brought to a shorter focus than that which passes through the center. The
result of this is that the image of a point is not a point but a "blur circle."
CHROMATIC ABERRATION occurs when light of various colors is re-
fracted by a lens made of one material. The light of shorter wavelength is
refracted more than that of longer wavelength, i.e., violet is brought to a
shorter focus than red. The image of a white point is a colored blur circle.
350 400 450 500 550 600 650
B G Y O R
WAVELENGTH
700
CHROMATIC ABERRATION of the human eye is corrected by various
stratagems which withdraw the cones from the region of maximum aberra-
tion, i.e., the shorter wavelengths. The horizontal coordinate of this diagram
is wavelength in millimicrons; the colors are indicated by initial letters.
the plate by laying on new light-sensitive
material, while simultaneously erasing
the old image."
Kiihne saw at once that with this pig-
ment which was bleached by hght it
might be possible to take a picture with
the living eye. He set about devising
methods for carrying out such a process,
and succeeded after many discouraging
failures. He called the process optogra-
phy and its products optograms.
One of Kiihne's early optograms was
made as follows. An albino rabbit was
fastened with its head facing a barred
window. From this position the rabbit
could see only a gray and clouded sky.
The animal's head was covered for sev-
eral minutes with a cloth to adapt its eyes
to the dark, that is to let rhodopsin ac-
cumulate in its rods. Then the animal
was exposed for three minutes to the
light. It was immediately decapitated,
the eye removed and cut open along the
equator, and the rear half of the eyeball
containing the retina laid in a solution
of alum for fixation. The next day Kiihne
saw, printed upon the retina in bleached
and unaltered rhodopsin, a picture of the
window with the clear pattern of its bars.
I remember reading as a boy a detec-
tive story in which at one point the de-
tective enters a dimly lighted room, on
the floor of which a corpse is lying.
Working carefully in the semidarkness,
the detective raises one eyelid of the
victim and snaps a picture of the open
eye. Upon developing this in his dark-
room he finds that he has an optogram
of the last scene viewed by the victim,
including of course an excellent likeness
of the murderer. So far as I know
Kiihne's optograms mark the closest ap-
proach to fulfilling this legend.
The legend itself has nonetheless
flourished for more than 60 years, and
all of my readers have probably seen or
heard some version of it. It began with
Kiihne's first intimation that the eye re-
sembles a photographic workshop, even
before he had succeeded in producing
his first primitive optogram, and it
spread rapidly over the entire world. In
the paper that announces his first suc-
cess in optography, Kiihne refers to thi:
story with some bitterness. He says: "1
disregard all the journalistic potentiali-
ties of this subject, and willingly sur-
render it in advance to all the claims of
fancy-free coroners on both sides of the
ocean, for it certainly is not pleasant to
deal with a serious problem in such com-
pany. Much that I could say about this
had better be suppressed, and turned
rather to the hope that no one will expect
from me any corroboration of announce-
ments that have not been authorized
with my name."
Despite these admirable sentiments
we find Kiihne shortly afterward en-
gaged in a curious adventure. In the
nearby town of Bruchsal on November
16, 1880, a young man was beheaded by
95
guillotine. Kiihne had made arrange-
ments to receive the corpse. He had
prepared a dimly lighted room screened
with red and yellow glass to keep any
rhodopsin left in the eyes from bleach-
ing further. Ten minutes after the knife
had fallen he obtained the whole retina
from the left eye, and had the satisfac-
tion of seeing and showing to several
colleagues a sharply demarcated opto-
gram printed upon its surface. Kiihne's
drawing of it is reproduced at the bot-
tom of the next page. To my knowledge
it is the only human optogram on record.
Kiihne went to great pains to deter-
mine what this optogram represented.
He says: "A search for the object which
served as source for this optogram re-
mained fruitless, in spite of a thorough
inventory of all the surroundings and
reports from many witnesses. The delin-
quent had spent the night awake by the
light of a tallow candle; he had slept
human eye as did the original subject of
the picture.
How the human eye resolves colors is
not known. Normal human color vision
seems to be compounded of three kinds
of responses; we therefore speak of it as
trichromatic or three-color vision. The
three kinds of response call for at least
three kinds of cone differing from one
another in their sensitivity to the various
regions of the spectrum. We can only
guess at what regulates these differences.
The simplest assumption is that the hu-
man cones contain three different light-
sensitive pigments, but this is still a
matter of surmise.
There exist retinas, however, in which
one can approach the problem of color
vision more directly. The eyes of certain
turtles and of certain birds such as chick-
ens and pigeons contain a great predomi-
nance of cones. Since cones are the or-
gans of vision in bright light as well as
RETINAL PHOTOGRAPH, or an optogram, was drawn in 1878 by the
German investigator Willy Kiihne. He had exposed the eye of a living rabbit
to a barred window, killed the rabbit, removed its retina and fixed it in alum.
from four to five o'clock in the morn-
ing; and had read and written, first by
candlelight until dawn, then by feeble
daylight until eight o'clock. When he
emerged in the open, the sun came out
for an instant, according to a reliable
observer, and the sky became somewhat
brighter during the seven minutes prior
to the bandaging of his eyes and his exe-
cution, which followed immediately.
The delinquent, however, raised his eyes
only rarely."
Color
One of the triumphs of modern pho-
tography is its success in recording color.
For this it is necessary not only to graft
some system of color differentiation and
rendition upon the photographic proc-
ess; the finished product must then ful-
fill the very exacting requirement that it
excite the same sensations of color in the
of color vision, these animals necessarily
function only at high light intensities.
They are permanently night-blind, due
to a poverty or complete absence of rods.
It is for this reason that chickens must
roost at sundown.
In the cones of these animals we find
a system of brilliantly colored oil glob-
ules, one in each cone. The globule is
situated at the joint between the inner
and outer segments of the cone, so that
light must pass through it just before en-
tering the light-sensitive element. The
globules therefore lie in the cones in the
position of little individual color filters.
One has only to remove the retina
from a chicken or a turtle and spread it
on the stage of a microscope to tee that
the globules are of three colors: red,
orange and greenish yellow. It was sug-
gested many years ago that they provide
the basis of color differentiation in the
animals that possess them.
In a paper published in 1907 the Ger-
man ophthalmologist Siegfried Garten
remarked that he was led by such retinal
color filters to invent a system of color
photography based upon the same prin-
ciple. This might have been the first in-
stance in which an eye had directly in-
spired a development in photography.
Unfortunately, however, in 1906 the
French chemist Louis Lumiere, appar-
ently without benefit of chicken retinas,
had brought out his autochrome process
for color photography based upon ex-
actly this principle.
To make his autochrome plates Lu-
miere used suspensions of starch grains
from rice, which he dyed red, green and
blue. These were mixed in roughly equal
proportions, and the mixture was strewn
over the surface of an ordinary photo-
graphic plate. The granules were then
squashed flat and the interstices were
filled with particles of carbon. Each dyed
granule served as a color filter for the
patch of silver-bromide emulsion that lay
just under it.
Just as the autochrome plate can ac-
complish color photography with a single
light-sensitive substance, so the cones of
the chicken retina should require no
more than one light-sensitive pigment.
We extracted such a pigment from the
chicken retina in 1937. It is violet in
color, and has therefore been named
iodopsin from ion, the Greek word for
violet. All three pigments of the colored
oil globules have also been isolated and
crystallized. Like the pigment of the hu-
man macula, they are all carotenoids: a
greenish-yellow carotene; the golden
mixture of xanthophylls found in chicken
egg yolk; and red astaxanthin, the pig-
ment of the boiled lobster.
Controversy thrives on ignorance, and
we have had many years of disputation
regarding the number of kinds of cone
concerned in human color vision. Manv
investigators prefer three, some four, and
at least one of my English colleagues
seven. I myself incline toward three. It
is a good number, and sufficient unto the
day.
The appearance of three colors of oil
globule in the cones of birds and turtles
might be thought to provide strong sup-
port for trichromatic theories of color
vision. The trouble is that these retinas
do in fact contain a fourth class of glob-
ule which is colorless. Colorless globules
have all the effect of a fourth color; there
is no doubt that if we include them, bird
and turtle retinas possess the basis for
four-color vision.
Latent Images
Recent experiments have exposed a
wholly unexpected parallel between vi-
sion and photography. Many years ago
Kiihne showed that rhodopsin can be ex-
tracted from the retinal rods into clear
water solution. When such solutions are
96
Eye and Camera
/^~~
Hg^
Mtt
11
Wv^
W
?rr ■•"
W"
p
1
'Hi
i1'
'n i
i;ui
!(■•'
.
FROG OPTOGRAM showing a
haired pattern was made hy the Gei-
nian ophthalmologist Siegfried Gar-
ten. The retina is mounted on a rod.
HUMAN OPTOGRAM was drawn
hy Kiiluie after he had removed the
retina of a beheaded criminal. Kit line
could not determine what it showed.
exposed to light, the rhodopsin bleaches
just as it does in the retina.
It has been known for some time that
the bleaching of rhodopsin in solution is
not entirely accomplished by light. It is
started bv light, but then goes on in
the dark for as long as an hour at room
temperature. Bleaching is therefore a
composite process. It is ushered in by a
light reaction that converts rhodopsin to
a highly unstable product; this then
decomposes by ordinary chemical reac-
tions—"dark" reactions in the sense that
they do not require light.
Since great interest attaches to the
initial unstable product of the light re-
action, many attempts were made in our
laboratory and at other laboratories to
seize upon this substance and learn its
properties. It has such a fleeting exist-
ence, however, that for some time noth-
ing satisfactory was achieved.
In 1941, however, two English work-
ers, E. E. Broda and C. F. Goodeve, suc-
ceeded in isolating the light reaction by
irradiating rhodopsin solutions at about
—73 degrees Celsius, roughly the tem-
perature of dry ice. In such extreme cold,
light reactions are unhindered, but or-
dinary dark processes cannot occur.
Broda and Goodeve found that an ex-
haustive exposure of rhodopsin to light
under these conditions produced only a
very small change in its color, so small
that though it could be measured one
might not have been certain merely by
looking at these solutions that any
change had occurred at all. Yet the light
reaction had been completed, and when
such solutions were allowed to warm up
to room temperature they bleached in
the dark. We have recently repeated
such experiments in our laboratory. With
some differences which need not be dis-
cussed, the results were qualitatively as
the English workers had described them.
These observations led us to re-exam-
ine certain early experiments of Kiihne's.
Kiihne had found that if the retina of a
frog or rabbit was thoroughly dried over
sulfuric acid, it could be exposed even to
brilliant sunlight for long periods with-
out bleaching. Kiihne concluded that dry
rhodopsin is not affected by light, and
this has been the common understanding
of workers in the field of vision ever
since.
It occurred to us, however, that dry
rhodopsin, like extremely cold rhodop-
sin. might undergo the light reaction,
though with such small change in color
as to have escaped notice. To test this
possibility we prepared films of rhodop-
sin in gelatin, which could be dried
thoroughly and were of a quality that
permitted making accurate measure-
ments of their color transmission through-
out the spectrum.
We found that when dry gelatin films
of rhodopsin are exposed to light, the
same change occurs as in very cold rho-
dopsin. The color is altered, but so
slightly as easily to escape visual obser-
vation. In any case the change cannot
be described as bleaching; if anything
the color is a little intensified. Yet the
light reaction is complete; if such ex-
posed films are merely wetted with wa-
ter, they bleach in the dark.
We have therefore two procedures-
cooling to very low temperatures and
removal of water— that clearly separate
the light from the dark reactions in the
bleaching of rhodopsin. Which of these
reactions is responsible for stimulating
rod vision? One cannot yet be certain,
yet the response of the rods to light oc-
curs so rapidly that only the light reac-
tion seems fast enough to account for it.
What has been said, however, has a
further consequence that brings it into
direct relation with photography. Every-
one knows that the photographic process
also is divided into light and dark com-
ponents. The result of exposing a film
to light is usually invisible, a so-called
"latent image." It is what later occurs in
the darkroom, the dark reaction of de-
velopment, that brings out the picture.
This now appears to be exactly what
happens in vision. Here as in photog-
raphy light produces an almost invisible
result, a latent image, and this indeed is
probably the process upon which retinal
excitation depends. The visible loss of
rhodopsin's color, its bleaching, is the re-
sult of subsequent dark reactions, of
"development."
One can scarcely have notions like
this without wanting to make a picture
with a rhodopsin film; and we have been
tempted into making one very crude
rhodopsin photograph. Its subject is not
exciting— only a row of black and white
stripes— but we show it at the right for
what interest it may have as the first
such photograph. What is important is
that it was made in typically photo-
graphic stages. The dry rhodopsin film
was first exposed to light, producing a
latent image. It was then developed in
the dark by wetting. It then had to be
fixed; and, though better ways are
known, we fixed this photograph simply
by redrying it. Since irradiated rhodop-
sin bleaches rather than blackens on de-
velopment, the immediate result is a
positive.
Photography with rhodopsin is onlv
in its first crude stages, perhaps at the
level that photography with silver
bromide reached almost a century ago.
I doubt that it has a future as a practi-
cal process. For us its primary interest
is to pose certain problems in visual
chemistry in a provocative form. It does,
however, also add another chapter to the
mingled histories of eye and camera.
97
RHODOPSIN PHOTOGRAPH was made by the author
and his associates Paul K. Brown and Oscar Starobin.
Rhodopsin, the light-sensitive red pigment of rod vision,
had been extracted from cattle retinas, mixed with
gelatin and spread on celluloid. This was then dried and
exposed to a pattern made up of black and white stripes.
When the film was wetted in the dark with hydroxyhv
mine, the rhodopsin bleached in the same pattern.
98
A device that operates on the principles of optics and
molecular physics— and that has an astonishing range
of applications.
The Laser — What it is and Does
J. M. Carroll
A chapter from a popular book, 1964.
Introduction
In i960, electronics scientists and engineers began to see
things in a different light.
It was a rich ruby light: not "kindled in the vine," as
the Persian poet Omar Khayyam said, but emitted by the
atoms of a synthetic gem stone.
The light came from the laser, a new device with wide
potential application in science, medicine, industry, and
national defense.
what's in a name?
The word laser is an acronym, or a word made up of the
first letters of several other words. Laser stands for Light
Amplification by Stimulated Emission of Radiation. It was
coined by analogy with another acronym: maser. Maser
stands for Microwave Amplification by Stimulated Emis-
sion of Radiation.
The maser works on the same basic principle as the laser
but, of course, emits microwave energy rather than light.
Masers are used as input amplifiers ( preamplifiers ) of radio
telescopes and space-tracking receivers that magnify fee-
ble signals gleaned from outer space.
No one is completely satisfied with the name "laser"
because lasers do not really amplify light in a strict sense;
99
instead they generate light with particular characteristics
that engineers and scientists find useful. In electronics
terminology a device that generates radiation is called
an oscillator, not an amplifier.
Furthermore, most lasers do not emit visible light at all
but rather infrared, or invisible, light. It is conceivable
that devices working on the same principle as the laser and
maser may someday emit ultraviolet or so-called black
light, X rays, or even gamma rays.
Scientists who moved from maser research into laser
research insist on calling the laser an optical maser. But
it can be argued that it is ridiculous to talk of "optical
microwave amplification by stimulated emission of radia-
tion" since "optical" means one thing and "microwave"
quite another.
Proponents of the term "optical maser" counter by say-
ing that "maser" doesn't stand for microwave amplification
by stimulated emission of radiation at all, but rather for
molecular amplification by stimulated emission of radia-
tion.
To the comment that masers do not amplify molecules
comes the answer that they depend for their action on the
behavior of the molecules of the substance.
Well, some masers and lasers do depend on molecular
effects. But more depend on the behavior of submolecular
particles: atoms, ions (atoms that have lost one or more
electrons), perhaps even electrons themselves.
Recently the term quantum device has been applied
to both masers and lasers, and this seems to make sense,
since the action of both the laser and the maser can be
explained by the science of quantum mechanics. In fact,
some scientists and engineers interested in lasers and
masers are attempting to form within the Institute of Elec-
trical and Electronics Engineers a professional group on
TOO
The Laser — What it is and Does
quantum electronics. And though for the present the term
"laser" seems deeply ingrained in the scientific vocabulary,
let's remember that the science we call electronics was
once known as thermionic engineering!
what's special about a laser?
The important thing about laser light is that it is co-
herent. The individual light rays are all of the same wave-
length or color, and are all in step. A laser beam differs
from a beam of ordinary light in both character and ef-
fectiveness in the same way that a platoon of well-drilled
soldiers differs from a ragtag, disorganized mob.
When light waves from a laser march in step, they can
perform amazing feats. The reason is that their energy is
not dissipated as the beam spreads out. This makes for an
intense concentration of energy at a very sharply defined
point. It also greatly extends the range of a light source.
Three of the many spectacular achievements of the laser
demonstrate how the properties of coherent light can be
put to work:
• Because its light does not spread out even at great
distances, a laser can illuminate the surface of the moon
with a two-mile- wide circle of light.
Laser beam on moon (black dot)
compared with area of radar
beam (shaded area) (Raytheon)
101
• Because its energy is concentrated at a fine point, it
can send a short, searing pinpoint of light into the human
eyeball to weld a detached retina back into place and re-
store sight.
• And since its radiation is so intense, it can burn holes
in a steel plate Ys inch thick at a distance of several feet.
These abilities have given rise to a whole range of ap-
plications. Laser range finders are used both by artillery
officers to sight their guns and by surveyors. In outer space,
where there is no atmosphere to absorb the light, the
laser will supplement conventional radar and radio for
space-vehicle navigation and communications.
Lasers can cut metal and other materials. But it is highly
unlikely that a laser will ever replace an engine lathe or
an oxyacetelene torch in most machining and metal-cut-
ting operations. Lasers are being used in the precision
machining of metals and in machining brittle materials
such as diamonds.
A laser can weld metals as well as retinas. But here, too,
its use is for precise work, as in making microelectronic
circuits. Nevertheless, large lasers mounted atop high
mountain peaks are being developed to provide a defense
against intercontinental-ballistic-missile warheads.
To the scientist, the laser is already a valuable tool in
absorption spectroscopy or the identification of compounds
by the particular wavelengths of light that they absorb.
Radiant Energy
How can a beam of light burn a hole in a steel plate?
It can do so because light is a form of radiant energy, and a
laser concentrates much radiant energy in a very tiny
spot. Radiant energy exists in many forms besides visible
light. It exists as radio waves, ultraviolet and infrared
light, X rays, gamma rays, and even cosmic rays.
102
The Laser — What it is and Does
WAVELENGTH AND FREQUENCY
It is sometimes convenient to think of radiant energy
as waves, that is, electromagnetic waves. Then the differ-
ent forms of radiant energy can be classified by their
wavelengths and arranged according to wavelength in a
spectrum. We have all seen the waves made by a pebble
thrown into a quiet pond. They are a series of alternating
crests and troughs. The wavelength is defined as the dis-
tance between two adjacent crests or two adjacent troughs.
Now, when a wave goes from crest to trough and back
to crest again, it is said to have gone through one cycle,
or alternation. The number of cycles that a wave executes
in one second is known as the frequency of the wave.
Light waves and all other electromagnetic waves travel
at the same speed, which is 186,000 miles, or 300,000,000
meters, a second. All scientific measurements are made in
metric system. In the metric system the basic unit of length
is the meter— a little over three feet.
RADIO SPECTRUM
The alternating current supplied by the power com-
pany is an electromagnetic wave that executes 60 cycles a
second; thus, in 1/60 of a second, or the time of one alter-
nation, the wave will travel 300,000,000/60, or 5,000,000
meters— roughly the distance from New York to Los An-
geles. The electromagnetic spectrum arranges the differ-
ent kinds of electromagnetic energy according to decreas-
ing wavelength.
Everyone is familiar with the red, orange, yellow, green,
blue, and violet spectrum of the rainbow after a spring
shower. The same separation of white light into its color
components occurs when we pass light through a glass
prism. A spectrum arranges the frequency components of
white light according to decreasing wavelengths. Similar
103
ELECTROMAGNETIC SPECTRUM
X-RAYS VISIBLE INFRARED
Jil^^
MICROWAVES RADIO
RUBY LASER
( ) lipS^
RUBY MASER
Electromagnetic spectrum from radio frequencies to X rays (Hughes)
spectra exist in the infrared and ultraviolet regions, but we
can't see them. They can, however, be photographed by
using special film. Radio waves also form part of the
electromagnetic spectrum.
A radio broadcasting station with a frequency of 1,000
kilocycles per second (or cycles per second times 1,000)
has a wave 300 meters long. A radar set used for naviga-
tion at sea has a wavelength of about 10 centimeters (one
centimeter equals 1/100 meter), or approximately 4 inches.
VISIBLE SPECTRUM AND INFRARED
Radiant energy is invisible to the human eye only until
we get to a wavelength of 0.00000075 meter, which we
see as red light. Since the meter is an ungainly unit for
measuring wavelengths of light, physicists use what is
called the angstrom unit, abbreviated A. One angstrom
equals 1/10,000,000,000 meter. Therefore we can say the
visible spectrum extends from 7,500 A ( deep red ) to 4,000
A, or blue. In between are regions of orange (about 6,000
A), yellow (about 5,900 A), and green (about 5,300
A).
The visible spectrum is bounded by longer waves of
104
The Laser — What it is and Does
infrared that we sense as heat. For example, a jet engine
exhaust has a wavelength of 40,000 A, while the heat of
the human body has a wavelength of about 99,000 A.
FROM SUN TANS TO COSMIC RAYS
The short wavelength, or blue end of the spectrum, is
bounded by the ultraviolet region. Sun-tanning ultraviolet
rays have a wavelength of about 3,000 A. Still shorter
are X rays ( 150 to 10 A) and gamma rays (1.4 to 0.1 A).
Gamma rays are associated with nuclear reactions, and
account for some of the deadly effects of atomic and hy-
drogen bombs and of radioactive waste materials. At the
high end of the spectrum are cosmic rays (0.01 to 0.001
A), those weird visitors from outer space whose effects
(they can cause biological mutations) are awesome in-
deed but about which very little is understood.
Scientists have known for a long time that the energy
of radiation is proportional to its frequency. We cannot
sense the presence of radio waves even though we stand
close by the antenna of a powerful broadcasting station.
Yet if we put a hand in front of a radar antenna, we may
feel a slight sensation of warmth. The energy of ultraviolet
waves will become painfully evident to some who sun
bathe not wisely but too well. The penetrating power of
X rays and gamma rays makes them useful for making
shadowgraphs of the human skeleton and internal organs
for medical diagnosis and for inspecting manufactured
parts for hidden flaws. Indeed, hard, or short, X rays and
gamma rays are used to destroy malignant tissue in the
treatment of cancer and related diseases.
The energy of each wavelet of radiation is called a
"quantum." It is measured by the frequency of the radia-
tion multiplied by Planck's constant (this is equal to
6.625 X 10~27 erg seconds— 26 zeros in front of the first
105
6). The intensity of a source of radiation depends upon
the number of quanta emitted from it that pass a desig-
nated boundary at a given time.
FLUORESCENCE
The action of the laser is allied to another, more familiar,
phenomenon, that of fluorescence. Fluorescence is said
to occur when radiant energy hits the atoms or molecules
of some particular material and in turn causes that sub-
stance to emit further radiant energy. Fluorescence has
this important property: the emitted radiation is always at
a lower frequency (longer wavelength) than the initial
radiation.
Here's how scientists explain fluorescence: Every atom
and molecule has certain energy states that it can occupy.
When the atoms absorb energy, they move to higher en-
ergy states. Conversely, when they return to lower energy
states, they give up energy, or emit radiation.
Imagine an atom to be a coil spring. When there is no
compression on the spring, it is in its ground, or rest, state.
When you compress the spring, you add potential energy
to the system. When you release the spring, it bounces
back and vibrates, giving up what is called its kinetic
energy.
In the picture tube of your television set, electrons
bombard a phosphor screen on the back of the faceplate.
The kinetic energy, or energy of motion of the rapidly
moving electrons, excites the atoms of the phosphor. As
these atoms relax, the faceplate of the picture tube glows,
and you see the television program because of fluorescence.
When a radiologist examines you with a fluoroscope, X
rays penetrate your body and excite the atoms of a phos-
phor screen. As the atoms of the phosphor coating relax,
the fluoroscope screen glows green, producing a shadow-
graph of the part of the body being visualized.
106
The Laser — What it is and Does
In a neon sign, an alternating current creates an elec-
tromagnetic field that agitates the molecules of neon gas
filling the tube. Because collisions of rapidly moving neon
molecules raise these molecules to higher energy levels,
they relax, emitting the orange-red glow characteristic of
a neon sign.
Of course, the common fluorescent lamp works on the
same principle of energy exchange. The inner walls of the
lamp tube are coated with beryllium oxide. Inside the
tube, there is an intense arc discharge between electrodes
at either end of the lamp tube. This arc discharge is rich
in ultraviolet light that energizes the phosphor molecules.
As these molecules relax, the lamp emits a blue-white light
similar to natural daylight.
We now have seen several examples of quantum energy
exchanges, but no one ever burned a hole in a steel plate
or illuminated the moon with a neon sign or with a fluores-
cent lamp. What, then, does the laser have that its less
powerful cousins lack?
Frequency Coherence
The answer is: the laser's coherence. In all the previous
examples of the phenomenon of fluorescence, the emitted
radiation had a broad spectrum. Because it was emitted
in random fashion, some wavelets added together while
others opposed each other.
Frequency coherence makes a big difference. It means
that all the emitted energy has the same wavelength. When
this happens, you can have a useful output indeed. Take
the babble of voices at a cocktail party as an example
of incoherent sound. The sound doesn't carry very far and
it is not especially meaningful. But if you were to concen-
trate all that sound energy into the blast of a police whistle
or siren, you could awaken half a city.
107
(WV
time
time
Frequency coherent radiation, top, and frequency incoherent radia-
tion, bottom (Raytheon)
Engineers learned many years ago that they could com-
municate more efficiently and more meaningfully when
they concentrated all the output of a radio transmitter at
a single frequency. But frequency coherence has other ad-
vantages besides efficiency. A beam of coherent light can
be modulated much as a radio signal can be. Modulation
is a process by which intelligence such as music or speech
is impressed upon a so-called carrier signal such as a radio
wave.
An incoherent light beam can be modulated in only the
most elementary manner— such as by switching it on and
off, as with the visual blinker lights used to send Morse
code between ships. But the frequency-coherent laser
beam can be modulated by such complex signals as speech,
music, or even a television picture.
Frequency-coherent light also lends itself to frequency
multiplication, the technique whereby a closely controlled
but relatively low radio frequency can be raised to a
higher output frequency. The output of a rubv laser at
108
The Laser — What it is and Does
6,943 A has been doubled to 3,472 A. The input was deep
red and the output blue-violet, almost ultraviolet. The
reason the wavelengths of laser light are given so pre-
cisely is that the emission of laser light depends on the
shifting of electrons between atomic orbits, and each
wavelength is characteristic of one particular orbital shift,
or so-called quantum jump.
Laser beams can also be mixed. For example, a ruby
laser operates in two slightly different modes. These modes
can be mixed in a microwave phototube. The frequency
difference between the modes yields a microwave signal
that can be handled by conventional radio or television
techniques. This property has permitted some engineers
to modulate laser beams with television pictures and to
recover the television signal after transmission for several
feet.
Scientists find the frequency coherence of the laser
especially gratifying. Before the discovery of the laser,
only signals in the lower, or radio, end of the spectrum
could be produced coherently. Radio techniques were
limited to producing signals whose wavelength was on the
order of a millimeter or so.
If monochromatic (or single-frequency) signals were
desired anywhere else in the spectrum, they had to be
produced by placing an appropriate filter in front of an
incoherent source. This method was unsatisfactory for
two reasons: it was very inefficient, since the source had to
produce many times the energy that could be usefully
employed; and, second, since no filtered output is ever
truly coherent, modulation, frequency multiplication, and
mixing were always unsatisfactory. But now a whole new
section of the spectrum, ranging from the "near" (to
visible light, that is) infrared to near ultraviolet, is open
to investigation, and there is evidence that the existing
109
gaps at the high and low ends of this laser operating range
can be filled by using related techniques.
SPATIAL COHERENCE
Frequency coherence is only part of the picture. The
output of a laser is also spatially coherent. This means that
all wavelets start in step with each other. Spatial coher-
ence also adds to the efficiency of a device. The difference
Spatially coherent radiation, top, and spatially incoherent radiation,
bottom (Raytheon)
between spatial incoherence and spatial coherence is like
the difference between a disorganized group of castaways
of a raft each paddling in his own way and the smooth,
efficient performance of a well-trained crew rowing an
eight-oared racing shell.
Ruby Lasers
The ruby laser was the first device to generate coherent
light successfully. The rubies used in lasers are synthetic
gem stones. They are made by fusing aluminum and
110
The Laser — What it is and Does
chromium oxides to produce large crystals. The amount
of chromium in a synthetic ruby is small— about five hun-
dredths of 1 percent. But it is that chromium upon which
laser action depends.
The ruby crystal is cylindrical, about )i inch in diameter
and i/2 to 2 inches long. It appears pink to the eye. That
is because there are two absorption bands in a ruby— one
at 5,600 A and the other at 4,100 A— which means that
when you hold a ruby up to the light, yellow-green light
and blue light are absorbed. This subtraction of yellow,
green, and blue from white light (which is a mixture of
all colors) gives the remaining light transmitted to the
eye its distinctive pink hue. Actually, there is also some
natural fluorescence in a ruby, but it is all but imper-
ceptible to the eye.
A laser crystal must be polished to optical flatness on
both ends. Both ends are also silvered, one with a heavy
coat while the other, or output end, is lightly silvered
with a coat that permits it to reflect only about 92 percent
of light incident on it.
Exploded view of ruby laser showing ruby, mirrors, and helical
flashtube (Hughes)
SILVER
MIRROR
SILVER
MIRROR
RED LIGHT WAVES TRAPPED BETWEEN MIRRORS
111
The ruby rod is now placed within a helical-shaped
xenon flashtube, the kind of tube widely used in elec-
tronic flash attachments for cameras. The process of ir-
radiating the ruby rod with a xenon flashtube is called
optical pumping. The output of the flash lamp is rich
in the yellow-green region.
The energy level of an atom ( an ion is just an atom that
has lost one or more electrons) depends upon the condi-
tion of its electrons. Now, an atom is like a miniature solar
system. It has a positive nucleus at its center in place of
the sun, and a specific number of planet-like electrons.
These electrons revolve around the nucleus and spin on
their own axes. Unlike the planets of the solar system,
however, each electron can occupy not just one but sev-
eral orbits. Moreover, the electrons can revolve around the
nucleus with different azimuthal momenta (speed) and
even change their direction of spin. Each change in orbit,
momentum, or spin corresponds to a discrete energy
level.
For example, when energy is imparted to an atom, an
electron may move to an orbit more remote from the
nucleus. The atom is said to absorb energy and to have
been raised to a higher or more excited energy state or
level. If the electron then returns to its original orbit, the
atom gives up energy; it may now emit light of a certain
precise wavelength. The atom is said to relax to a lower
or less excited energy state or level. When light wavelets,
or photons, at 5,600 A from the flashtube irradiate the
ruby rod, they raise the energy of some of the chromium
ions dissolved in the ruby from ground state (1) to various
levels lying within the absorption band. Then the chromium
ions immediately begin to drop from these higher energy
levels. Some drop right back to the ground state— level (T)
—as they do in natural fluorescence. But others drop to
112
The Laser — What it is and Does
20 r
5 15
rO
O
? 10
o
uj 5
■- "A
y
i"M
_L_L.
.1 I I I I I I I I I 1 1
6,925 R2
"I
WAVELENGTH IN A
6,950
6,925 R2
Hi i ■ i i ■ I
WAVELENGTH
IN A
6,950
(B)
(0
Energy level transitions in a ruby laser as described in text (A); low-
level pumping (B) and high-level pumping (C) showing how latter
mode concentrates energy at one wavelength (Electronics)
an intermediate or so-called metastable state ©. If left
alone, the latter chromium ions would continue their drop
to level ©, and the result would just be natural fluores-
cence. But these ions dally for a short but measurable
time in level ©, and this is what makes laser action
possible.
While the chromium ions are trying to get back to level
©, the flashtube keeps on irradiating more chromium ions.
113
In fact, the two-step movement from state (T) to state @
and down to state (2) is much faster than the movement
from state (2) to state (7). Thus there develops a chromium-
ion traffic jam at energy level (2).
STIMULATED EMISSION
As the pile-up of chromium ions in level (2) continues,
another situation develops: soon there are more chromium
ions in level (2) than in level (1). This is called population
inversion, and is essential for laser action.
When you have inversion of the chromium ion popula-
tion, the laser resembles a spring that is wound up and
cocked. It needs a key to release it. This is what is meant
by stimulated emission of radiation: the stimulus is the
key that releases the cocked spring.
The key is a photon of light of exactly the wavelength
to be emitted (6,943 A). Emission begins when a random
chromium ion spontaneously falls from level (2) to level
(1) emitting a photon at 6,943 A. The photon strikes
neighboring metastable (level (2)) ions, causing them to
emit additional photons, and these in turn trigger other
metastable ions.
As the photons travel along the rod, some emerge from
the sides of the cylinder and are lost. Others hit the sil-
vered ends of the cylinder and are reflected back into
the rod. The reflections tend to favor those photons that
are traveling parallel to the long axis of the cylinder. And
so, there is now a stream of photons bouncing back and
forth between the silvered ends of the cylinder. The pho-
Two following pages: How a ruby laser works. Pumping light ir-
radiates ruby rod (A) raising some atoms to their metastable state
(B). One atom spontaneously emits coherent radiation (C) triggering
other nearby atoms (D). Photons emitted parallel to sides bounce
back and forth between mirrors triggering other atoms (E) until
light pulse (F) bursts from slightly transparent end (Electronics)
114
The Laser — What it is and Does
PUMPING LIGHT
A
V_J / /\ \ I / \
° I
nVV
m
SLIGHTLY-TRANSPARENT MIRROR
(A) RUBY ROD^
115
(D)
•
o
• o
o
o
•
•
D
A
O
? 5 ■
o
•
• o
O
X
(E)
•* — #-o ■* — *• # < — * o *—
•* — »■ o •* — *• • < — *■ o *-
D
— • —
-► o *-
A
— O —
B
— >
— ►
« — » o « * 0 *__* o «-
-* • *-
^ O —
C
— *•
0
(F)
400 800
TIME IN MSEC
1,200
116
The Laser — What it is and Does
tons become more numerous, and consequently the light
beam grows more intense as the photons already in the
stream trigger still more metastable chromium ions into
emitting their radiation.
Eventually the photon stream builds up sufficient in-
tensity so that it bursts from the partially silvered end of
the ruby as a single pulse of monochromatic ( single color
or frequency), spatially coherent light.
PARALLEL RAYS
The light beams coming out of the partially silvered
end of the ruby rod are almost exactly parallel, and it is
this factor that makes it possible for a laser beam to reach
the moon. Conventional light sources such as an in-
candescent lamp are point sources: their light rays are
emitted in a spherical pattern. Conventional rays can be
made parallel by use of focusing mirrors and lenses, but
such optical systems are far from efficient: the light beam
diverges, and consequently loses its intensity at great dis-
tances. But since the beams coming from a laser are
parallel to begin with, they remain essentially parallel
even at exceedingly great distances.
Liquid and Plastic Lasers
The ruby laser was the first laser, but today it is only
one member of the class of optically pumped lasers.
Furthermore, there are many varieties of ruby lasers. The
original ruby lasers worked at room temperature. Later
devices have been designed to work at cryogenic tempera-
tures, or temperatures close to absolute zero ( — 273 de-
grees centigrade). Cryogenic temperatures are usually
achieved by immersing the laser in liquid nitrogen or
liquid helium. Lasers cooled this way can put out a con-
tinuous beam of coherent light instead of a series of flashes.
117
Other optically pumped lasers include many different
crystalline materials, most of which are doped: made im-
pure by the infusion of small quantities of some other
material— either a rare-earth element, such as europium
or neodymium, or an actinide element— a class of heavy
metals that includes uranium. Some optically pumped
lasers have been made of doped glass ( glass to which im-
purities have been added), of liquid or gas in a quartz
cavity or of bundles of plastic fibers.
Gaseous Lasers
The gaseous laser represents a second general class of
laser. The working medium is a mixture of helium and
neon gas at very low pressure (o.i millimeter of mercurv
Helium-neon gas laser (Raytheon)
118
The Laser — What it is and Does
of neon and 1.0 millimeter of mercury of helium). The gas
is contained in a cylindrical Pyrex tube about one meter
long and 17 millimeters in diameter. At each end of the
tube is a quartz plate ground optically flat and with a
13-layer dielectric (or electrically nonconductive ) coat-
ing on its inner face: this coating produces the same effect
as the lightly silvered end of the ruby rod. The spacing of
the quartz-plate mirrors can be changed with precision
for optimum internal reflection, thanks to an arrangement
known as a Fabrtj-Perot interferometer. The laser beam is
emitted from both ends of the apparatus.
ELECTRICAL PUMPING
The gas laser is not optically pumped, nor is it pulsed
at the rate of three or four times a second as is the ruby
laser. Instead it operates in a continuous-wave mode, its
excitation supplied by a radio-frequency field— though in
some gas lasers, direct current has been used to produce
the required discharge. In a typical gas laser the source
is a 50-watt transmitter operating on a carrier frequency
of 29 megacycles per second. This frequency was selected
simply because it lies within a band provided by the Fed-
eral Communications Commission for industrial, scientific,
and medical use; another frequency would do equally well.
The transmitter is coupled to the gas tube by three metal
loops.
The radio-frequency generator produces an electrical
discharge through the gas that raises the helium gas atoms
to an excited state designated as the 23S state. This is a
metastable state that the helium atoms retain for a finite
period of time.
When the helium metastables collide with neon atoms
in the ground state, the helium atoms transfer their energy
to the neon atoms and drop immediately to the ground
119
25,
20
19
o
cr.
17
g '6
24.6
19.81
He'
23S
Ne +
oL
He
Energy levels in a helium-neon laser (Electronics)
state. Simultaneously, the neon atoms are raised to the
so-called 2S state because the energy level of the 2s state
in neon is nearly equal to the energy level of the 23S
state in helium.
There are three excited states in neon that are involved
in this reaction: the 2S, 2p, and is states. We are primarily
interested in the transition between the 2S (higher) and
2p (lower) states. The 2s state is a metastable state. Ac-
tually, there are four substates in the 2S band and ten
substates in the 2p band. Theoretically there are 30 possi-
ble transitions, or downward changes in energy level, that
could occur, with each giving off radiation at its character-
istic wavelength. Actually, only five of these transitions
have as yet figured importantly in stimulated emissions;
all correspond to wavelengths in the near-infrared region.
The strongest of these emissions is one at 11,530 A.
As in the case of the ruby laser, neon atoms tend to pile
120
The Laser — What it is and Does
up in the 2s state, and the threshold energy is the amount
of input energy that makes the population of neon atoms
in the 2S state equal to that in the 2p state. When some
random neon atom spontaneously makes the transition
from the 2s state to the 2p state, radiation at 11,530 A
stimulates coherent emission.
The photon at 11,530 A stimulates nearby metastable
neon atoms, and they, too, go down the chute and emit
their photons at the same wavelength. Photons emitted
perpendicular to the Fabry-Perot mirrors bounce back
and forth between the mirrors until they acquire sufficient
intensity to break out. Photons emitted in other direc-
tions are lost through the walls of the tube and do not
participate in coherent emission.
When in operation, a gas laser is bathed in an orange-
red glow, but this light has nothing to do with its laser
action. Most of the coherent output of the gas laser is in
the infrared region and is invisible to the eye. The visible
glow results from spontaneous transitions of excited neon
atoms that do not enter into the stimulated emission of
radiation. In fact, the glow is identical to that of any neon
sign.
Injection Lasers
The third basic type of laser is the injection laser. An
injection laser consists of a semiconductor diode made of
gallium arsenide or of gallium arsenide-phosphide.
A diode is an electronic part that has the property of
conducting current easily in one direction but almost not
at all in the opposite or reverse direction. The injection
laser is a forward-biased semiconductor diode. It conducts
current in its easy direction.
A semiconductor is a material that does not conduct
electricity so well as something like copper does, but does
121
so better than an insulator such as sulphur. The most com-
mon semiconductors are the metals silicon and germanium,
but some compounds can also be used, and, for the injec-
tion laser, gallium arsenide has proved useful. Because
gallium is a little better conductor than silicon, and arsenic
a little poorer, when mixed together they give roughly
the same effect as silicon.
Now, to make a diode out of a block of semiconductor
material, it is necessary to dope it. This is done by allow-
ing the two impurities— tellurium and zinc— to diffuse into
the block at high temperature. Because the tellurium atom
has one more valence (combining) electron than does
arsenic, when tellurium atoms replace some of the arsenic
atoms in the gallium-arsenic block, there are a few free
electrons left over. Since the electron has a negative
charge, tellurium-doped gallium arsenide is called N-type,
or negative, gallium arsenide.
Because zinc, on the other hand, has one less valence
electron than gallium, when some zinc atoms replace a
few of the gallium atoms, there are several holes, or elec-
tron deficiencies, left over. Therefore, zinc-doped gallium
arsenide is called P-type, or positive, gallium arsenide.
The boundary where the regions of N-type and P-type
gallium arsenide meet is called the semiconductor junc-
tion. If you connect the positive terminal of a battery or
electronic power supply to the P-type region of a semi-
conductor diode and connect the negative terminal to the
N-type region, the diode will be biased in the forward
direction, and current will flow easily across the semi-
conductor junction. If the power supply is connected with
its negative terminal going to the P-region and its positive
terminal going to the N-region, the diode will be biased
in its reverse direction, and little, if any, current will flow
across the semiconductor junction.
122
The Laser — What it is and Does
HOW DOES IT WORK?
Scientists are not yet sure just what energy transitions
occur in the injection laser. But laser action seems to be
most pronounced on the P-side of the junction. This might
indicate that some energetic electrons making up the cur-
rent flowing across the junction recombine with holes and
give up energy in the recombination process.
The injection laser emits coherent light by passing ex-
tremely high current between the terminals of the semi-
conductor diode, so that light is emitted along the line
that defines the semiconductor junction. The light comes
out incoherently at first, but as the intensity of the current
is increased, the emission becomes coherent. Of course, all
Semiconductor injection laser design as developed by IBM (Elec-
tronics)
INTENSITY
(ARBITRARY UNITS)
I50t
pTYPE GaAs
100
n TYPE GaAs
(ABOVE THRESHOLD
8.13 AMPS 50
Oo^ (AT THRESHOLD) 8.03 AMPS
70 60 50 40 30 20 10
ANGLE OF ROTATION (DEGREES)
123
this electrical current passing through the relatively small
diode makes the diode heat up rapidly. Since such extreme
heating could destroy the semiconductor junction, before
the diode is operated it is usually immersed in a cryostat,
or double bottle, the inner bottle filled with liquid helium
and the outer one with liquid nitrogen. Furthermore, the
current is usually pulsed rather than passed continuously.
A typical injection laser is a rectangular parallelopiped
(six-sided solid block whose opposite faces are parallel)
about ten times as long as it is wide. Dimensions of a
typical unit are 1/10 by 1/10 by lM millimeters. The sides
are finely polished and tend to reflect light back into the
laser so that the emission of coherent light comes out in
parallel rays from the square sides of the block. Silvering
is not required because the block itself is metallic, and
when its sides are polished they will reflect the light rays
generated within the block.
Current is applied to opposite rectangular sides of the
block. The current flow is perpendicular to the semicon-
ductor junction, which is a narrow plane or region cutting
the block along its long axis.
The reflection of waves at the polished sides of the diode
tends to favor the waves coming out of the square ends
parallel to the junction. Furthermore, since the recom-
bination process takes place all along the semiconductor
junction plane, coherent-light waves traveling along the
junction stimulate radiation from other hole-electron pairs,
and the wave grows in intensity before it bursts from the
square sides of the laser.
A gallium-arsenide laser emits coherent light at 8,400 A
in the near-infrared region. This light is invisible to the
human eye. Gallium arsenide-phosphide lasers have emit-
ted coherent light at 7,000 A, in the deep-red region.
Furthermore, by varying the amount of phosphorus in the
124
The Laser — What it is and Does
laser, the color can be changed throughout the near-
infrared and deep-red regions of the spectrum. Several
other intermetallic compounds involving indium and anti-
mony as well as gallium, arsenic, and phosphorus show
promise of producing laser action. A silicon-carbide diode
was reported to have emitted blue-violet light, but proof
of this accomplishment is as yet inconclusive.
The current passed through the particular laser we have
described may vary from 10 to 25 amperes or more. At
lower currents, the emission is incoherent and involves
only a small part of the junction area. As current is in-
creased, the area of incoherent sparkling or sporadic emis-
sion of light spreads out along the junction, and coherent
emission can be noticed near the center of the junction.
Comparison
Thus there are three main types of lasers: optically
pumped lasers, which may be crystalline, glass, liquid,
gaseous, or plastic; radio-frequency or direct-current-
pumped gas lasers; and semiconductor diode lasers
pumped by injection of high current.
GASEOUS LASERS
The gas laser emits coherent light, usually in the infra-
red region. Gas lasers are used mostly in scientific investi-
gations, such as spectroscopy, and for experiments in space
and time, such as verification of some of the consequences
of the theory of relativity. The gas laser is useful in these
investigations because its output is the most nearly co-
herent of all lasers and because continuous output is
conveniently available from gas lasers even at room tem-
perature.
Because gaseous lasers operate in the continuous wave
mode rather than through pulsation, they have proved
125
better than optically pumped lasers for many communi-
cations experiments, such as the transmission of speech
and music or television pictures.
Furthermore, since gas lasers produce the most nearly
coherent output of any laser— the only thing that can cause
a helium-neon gas laser to deviate from its 11,530 A center
frequency is mechanical vibration of the apparatus— they
have been used for scientific studies, such as checking the
experimental evidence of Einstein's theory of relativity
and for constructing a precise gyroscope.
OPTICALLY PUMPED LASERS
Optically pumped lasers are used when high energy is
required, such as for burning metal, performing delicate
eye operations, precision welding or machining. The most
used optically pumped laser is still the ruby laser. It is
one of the few lasers that can give visible output. Nearly
all gas lasers, and most types of optically pumped lasers,
work in the infrared region. Most optically pumped lasers
emit pulses at a relatively low repetition rate. Continuous
output can be achieved only by putting the laser in a
cryostat, or double bottle of liquid helium and nitrogen.
Although the physical form of a ruby laser is simpler than
that of a gas laser, its excitation system is somewhat more
complex. The gas laser needs only a simple radio trans-
mitter, while the ruby laser requires an electronic flashgun
and either a special xenon flashtube or a carefully designed
system of reflectors.
INJECTION LASERS
The injection laser is physically simpler than either the
ruby or gas laser. For excitation, it actually needs only a
rudimentary direct-current power supply, but it is usually
operated in a cryostat. Injection lasers can produce a
whole range of coherent output frequencies within the red
126
The Laser — What it is and Does
and infrared regions of the spectrum. They deliver con-
tinuous or nearly continuous output, and they, too, have
been found useful in communications experiments in
which speech, music, or even television pictures have been
transmitted. Gallium-arsenide diodes operated at lower
current and at room temperature are already being used
in portable communications systems. Although the infra-
red output of these devices is not coherent, they have
permitted communications over a range of thirty miles.
Universal Coherence
Sciences have long dreamed of generating coherent
emission at all frequencies of the electromagnetic spec-
trum. Quantum devices have made important contribu-
tions toward this end, but a great deal remains to be done.
It has been suggested that variations of the word "maser"
be coined for all the new devices, including the ones yet
to come. There might be rasers ( radio-frequency ) , masers
(microwave), irasers (infrared), lasers (Zight), uuasers
( ultraviolet ) , xasers (X ray), and gasers (gamma-ray).
One prominent scientist jocularly suggested the name
"daser," standing for "darkness amplification by stimu-
lated emission of radiation."
All this points up the advantage of talking about quan-
tum devices (and specifying whether they are oscillators,
amplifiers, or harmonic generators) and designating the
wavelength of interest rather than playing with acronyms.
It does, nevertheless, seem to be a fact of life that the term
"maser" will continue to be used both for amplifiers and
for oscillators not only in the microwave region (roughly
1,000 megacycles per second) but perhaps for devices
operating at even lower frequencies, when and if such de-
vices are developed.
Likewise, it seems that the term "laser" will continue to
127
be used to refer both to amplifiers and to oscillators that
operate in the near-infrared, visible, and near-ultraviolet
portions of the spectrum. Neither extension of laser action
into the far-infrared ( near microwaves ) nor into the far-
ultraviolet (near X rays) will result in a change in termi-
nology.
But possibly, when we can successfully generate co-
herent X rays and gamma rays, another term will be used,
for already, as mentioned above, the word "gaser" is being
bandied about.
MASERS
Masers are usually true amplifiers instead of the gen-
erators that lasers are. This means that they receive a
weak signal and pass it on at a higher power level. Masers
operate between 300 megacycles per second ( 100 centi-
meters or 1 meter wavelength) and 100,000 megacycles
per second (3 millimeters).
We might remark parenthetically that there is other
millimeter-wave research going on that does not involve
masers. One special microwave tube, the Tornadotron,
has been reported to have an output of 500,000 mega-
cycles per second, or a wavelength of 0.6 millimeter.
A typical maser consists of a crystal containing chro-
mium that is pumped by the output of a microwave tube
operating at a frequency much higher than the one to be
received. The microwave signal pumps the chromium ions
to an elevated energy level that is metastable.
Incoming signals at a certain lower microwave fre-
quency stimulate the chromium ions to fall from their ele-
vated energy level to an intermediate level before the
ground state. In so doing, they emit radiation at the fre-
quency of the incoming signal and thus amplify it.
To avoid the introduction of noise or unwanted signals,
128
The Laser — What it is and Does
maser amplifiers are placed between the pole pieces of a
powerful magnet, and are operated in a double bottle with
liquid helium on the inside and liquid nitrogen on the
outside.
About a dozen radio astronomical observatories
throughout the world use maser amplifiers to pick up
radio-frequency emissions from distant planets, stars, and
nebulae. Several stations use maser amplifiers for tracking
satellites and space probes. So do some of the stations that
receive radio and television signals from orbiting com-
munications satellites such as Telstar and Relay. It is
possible that maser amplifiers are used in special military
radar and communications applications, but if so, the
Department of Defense isn't saying!
INFRARED LASERS ( IRASERS )
Various kinds of lasers cover the near-infrared spectrum
from nearly 13,000 A right up to visible light. This leaves
a gap in the spectrum from 3 millimeters wavelength to
0.013 millimeter. This gap includes the millimeter and
submillimeter-wave regions of the radio spectrum and the
far-infrared band that encompasses radiation from warm
and lukewarm objects.
NEW COLORS IN LASERS
Progress has not been so good in the visible region.
Only a few lasers produce visible light, and most of that,
as we have noted, is deep red. There is, of course, the ruby
laser. Red light has been produced by several other meth-
ods as well: by a laser consisting of a crystal of calcium
fluoride with the rare-earth samarium dissolved in it; from
europium chelate (rhymes with "tea late") embedded in
a plastic tube (a chelate is a complex organic or hydro-
carbon molecule containing a metal atom, in this case an
atom of the rare-earth europium); with the gallium ar-
129
senide-phosphide laser; and with some helium-neon gas
lasers.
There is a demand for lasers to produce other colors
besides red. The Navy would like to have a blue-green
laser because blue-green light is best for penetrating sea-
water and because a blue-green laser could be used as
part of an underwater television system to help navigators
of nuclear submarines detect the presence of friendly or
hostile submarines or other underwater objects.
So far, the only progress in that direction has been the
development of "blue-violet lasers," produced by doubling
the output frequency of a deep-red laser. (Doubling the
output frequency is the same thing as dividing the wave-
length by two.) Likewise, there are "green lasers,"
achieved by doubling the output frequency of lasers op-
erating in the near-infrared region.
But when you double the output frequency of a laser,
you lose 8/10 or more of its energy, and what's left will
hardly perform the job the Navy has in mind. Therefore
the search for different colored lasers continues, with
scientists now studying not only rare-earth and actinide
metals but even various organic compounds. They feel
that, given the right conditions, any substance that will
fluoresce can be made to lase. This leaves them with thou-
sands of compounds to investigate.
ULTRAVIOLET LASERS ( UVASERS )
So far the story of the ultraviolet laser is short and
sweet. One optically pumped laser, using a glass rod in
which a small quantity of the rare-earth gadolinium has
been dissolved, lases at 3,125 A in the near ultraviolet.
GAMMA-RAY LASERS ( GASERS )
Nothing has been announced officially about X-ray
lasers, but certain work is going on with gamma-ray lasers
130
The Laser — What it is and Does
under Navy auspices, though the work has not progressed
very far as yet. The Russians have also announced work
in this field.
The approach is to use a gamma-ray-emitting isotope
of ruthenium to raise a radioactive isotope of rhodium to
a higher energy state that is metastable. After a half-life
of some 40 days, the level of energy emitted by the
ruthenium will drop to that of the metastable state of the
rhodium isotope, and trigger emission at roughly 0.3 A.
There are many problems in the way, however. First,
one has to find a way to make a crystal containing the
appropriate isotopes without changing their essential
characteristics. Next comes the problem of containing the
gamma rays (they will penetrate just about anything) so
as to achieve spatial coherence. If achieved, a gamma-ray
laser would be a death ray in every sense of the word.
Gamma rays have several times the burning power of
X rays, which are, of course, harmful when improperly
applied.
The Future of the Laser
As we have seen, the laser has the advantage of provid-
ing a monochromatic or single-color light source. Further-
more, its beam is so collimated that all its energy can be
focused on a very small spot. It is also highly directive,
with little or no tendency for the beam to bend or spread
out even over the astronomical distances of outer space.
These properties have suggested a great many uses in na-
tional defense, industry, medicine, and science.
Lasers may be developed into devastating antiperson-
nel weapons for use on the battlefield. They may be sent
into space on special platforms to fight intercontinental
ballistic missiles or to destroy hostile space stations or
satellites. The laser may also be used to modify chemical
131
compounds or even to change the genetic characteristics
of the protein molecules of living organisms.
Someday special fiber-optic light pipes or other optical
wave guides, such as evacuated tubes with an internal
mirror system, may carry laser signals much as coaxial
cables now carry telephone conversations and network
television programs between cities. A fiber-optic light pipe
is a very fine glass, plastic, or arsenic-trisulfide rod pol-
ished on the outside; its walls reflect light back inside so
that it can bend around corners and still carry a light
beam.
One way to put a TV signal on a laser beam is first to
impress the complete TV picture and sound (the video
signal) on a microwave carrier. The microwave carrier is
then used to excite a special crystal situated in a micro-
wave cavity or special metal box. When the laser beam
traverses the crystal, entering and leaving the cavity
through small side windows, the beam is modulated or
made to vary in accordance with the modulated micro-
wave signal. At the receiver, the beam of a microwave
traveling-wave amplifier phototube is similarly made to
vary in accordance with the variations of the laser light
striking the traveling wave tube's photocathode. We now
have again the microwave carrier with the video signal
riding on it. This signal is demodulated, using conven-
tional electronic circuits to give the original TV picture
and sound.
A wideband video channel can be divided into many
subchannels, actually some 600, each of which can carry
a telephone conversation. Electronic circuits called filters
slice up the video channel into so-called voice channels.
Each voice channel is about o to 2,000 cycles per second
wide. Each incoming telephone signal is heterodyned, or
moved up, in frequency to fit a specific voice channel at
132
The Laser — What it is and Does
the transmitting end, then moved down in frequency and
routed out on its proper telephone line at the receiving
end.
A laser communications system would greatly expand
the capabilities of our nationwide telecommunications
network. Tiny lasers may also function as parts of the
memory system of a computer. Such a computer would
literally work with the speed of light.
Who knows? You may even one day have a laser igni-
tion system in your automobile!
MILITARY USES
One of the first uses that occurs to most people is to
build a big, superpower laser and use it to shoot down
ballistic missile nose cones. This would, they reason, make
our nation secure from the terrors of thermonuclear war.
But it isn't as easy as all that. Even the most powerful
lasers can at present penetrate only /8-inch of high-carbon
( easily burnable ) steel. And the holes they make are mere
pinpricks. Furthermore, burning requires that the laser be
only a distance of a few feet from the steel. At longer
ranges, the water vapor and dust in the atmosphere se-
verely reduce the effective power of the light ray.
Nevertheless, the Air Force is hard at work trying to
develop big lasers and figuring out how to deploy them
effectively outside the earth's atmosphere: atop mountain
peaks, aboard orbiting satellites, or even on antimissile
missiles.
Meanwhile, the military and space agencies have other,
more prosaic, but none the less vital uses for the laser.
When the Apollo lunar capsule carries the first Americans
to the vicinity of the moon, the two-man crew aboard the
Lunar Excursion Module that will make the actual land-
ing on the moon will probably use a laser altimeter to feel
133
their way onto the lunar surface. Before that, astronauts
in Project Gemini will use laser radar to practice rendez-
vous and docking of satellites in space. Already a large
laser at Wallops Island, Virginia, has tracked an orbiting
satellite 1,000 miles up. Incidentally, at that range the
laser beam was only 200 feet in diameter.
The Army has ordered several laser range finders for
use on the battlefield. They will be able to measure the
distance to targets far more accurately than their optical
or radar counterparts.
During World War II the Army made effective use of
sniperscopes and snooperscopes, infrared devices that lo-
cated targets even at night. But for such devices to be
effective, the target had to be a good deal warmer than
the background. Now, with an infrared laser, it would be
possible to scan the target and get a picture regardless of
its temperature.
During World War II the Navy used infrared "Nancy"
equipment (usually Nerst tubes or hot filaments enclosed
by a black metal hood and placed behind a deep ruby
lens) for short-range communications between ships. But
the laser affords a much more efficient and less easily de-
tectable source of infrared.
The Armed Forces have a project under way to see just
how fast a computer can operate. Some people think that
the result will be a new high-speed giant brain for our
ballistic missile early-warning system. But a better guess
is that such a computer will be used to crack secret enemv
codes and ciphers. Anyway, one part of this project is a
laser computer, sponsored by the Air Force, in which light
pulses would do the counting instead of electrical signals.
Such a computer would be faster by several orders of
magnitude than any computer now available, since light
travels faster than electrical current, which is slowed
134
The Laser — What it is and Does
down by the action of reactive elements, such as capacitors
and inductors in the circuit.
INDUSTRIAL APPLICATIONS
Industry is already using lasers to perform delicate
machining and welding operations in the manufacture of
microelectronic circuits.
A microelectronic circuit is fabricated on a thin wafer
of silicon. Sometimes forty circuits are made at one time
on a wafer only an inch in diameter. Each circuit can do
the work of, say, a five-tube radio or perhaps a computer
stage.
The circuits are made by allowing certain selected im-
purities to diffuse into the silicon wafer in prescribed pat-
terns. These patterns are formed by first allowing a film
of silicon dioxide ( glass ) to grow over the silicon wafer—
usually by applying steam to the surface— then selectively
etching away portions of the film.
Selective removal of the oxide is accomplished by first
coating the oxide with so-called photoresist— a film that
becomes tough and acid-resistant when exposed to light-
then masking the wafer with a diffusion mask and expos-
ing the unprotected photoresist to light. The wafer is next
etched with strong acid, and its silicon-dioxide coat is
eaten away except where it is protected by light-hardened
photoresist.
Preparation of the diffusion mask is a critical operation,
and laser machining of metallic foil is expected to allow
making sharper and more precise pattern outlines. Possibly
lasers may be used to remove the oxide itself, thus saving
several steps in the process of manufacturing micro-
circuits.
Laser light sources could be valuable in high-speed
photography where chromatic aberration or the unequal
135
bending of light of different wavelengths through the
camera lens can cause a blurred image.
Since different components of the atmosphere absorb
different wavelengths of light to a greater or lesser extent,
a bank of lasers used at an airport as a transmissometer
could disclose not only the visibility at the end of the
runway— as the optical devices already in use do— but also
the makeup of the atmosphere at any particular time. Such
a laser device could also be useful in air-pollution studies.
( Transmissometers are used even though the end of the
runway may indeed be visible from the control tower; the
view from the tower is not what an approaching pilot sees;
besides, the instrument, unlike a human observer, remains
on duty around the clock. )
In a chemical process, a laser might be created so that
its beam is absorbed to a great extent by the desired prod-
uct. The laser could be focused permanently through the
output pipe, and automatic control equipment could be
adjusted so that the product absorbs maximum light from
the beam. This would assure that the product in the output
pipe has precisely the desired chemical composition.
The ability of a laser beam to carry an almost infinite
amount of information has set communications engineers
to speculating about its possible use for trunkline or inter-
city communications. Today, these are handled by coaxial
cables or microwave links. One microwave link can earn1
four television programs simultaneously or replace any
one of the television channels with up to 600 telephone
conversations. But a laser beam could carry many times
this amount of information.
Nevertheless, since dust and water vapor in the atmos-
phere severely reduce the effective power of a laser beam,
a serious problem still remains before lasers can be used
for practical communications. Of course, short-distance
136
The Laser — What it is and Does
communications would indeed be possible, as would com-
munications to and from communications satellites. In the
latter case, the beam travels in the earth's atmosphere for
only a relatively short distance, although the total trip
might be 1,000 miles or even more.
One answer to abetting laser communications would be
to use light pipes or evacuated tubes with mirrors arranged
to conduct the beam around corners where necessary.
A laser telephone exchange has been contemplated.
Here the light pulses would be conducted by fiber optic
strands. These strands carry light around comers just as
copper wires carry electrical current. Though a fiber-optic
strand severely cuts down the power of the light being
transmitted, in a telephone exchange the length of the
interconnecting strands can be kept short by design. The
big advantage of a laser telephone exchange would be
that there would be no crossed wires or unwanted pickup
between adjacent optical fibers so that you would not
occasionally hear fragments of another conversation on
your line.
MEDICAL USES
Lasers have been regarded as a major boon to medi-
cine. Thousands of Americans suffer each year from a de-
tached retina. In this condition the retina, the light or
sensitive area at the rear of the eye, comes loose from the
inner surface or choroid coating of the eyeball. The fluid,
or humor, with which the eye is filled works in behind
the retina and aggravates the condition. Initially, the con-
dition causes distorted vision, but if the retina becomes
completely loose from the optic nerve, blindness results.
A laser beam can be focused through the lens of the eye
so that it makes small scars around the periphery of the
retina and thus welds it back into place.
137
A laser can also burn out small tumors in the eye. In
fact, a laser beam can be made as narrow in diameter as
the diameter of a single human cell. Some surgeons see
the laser, therefore, as a device that can burn out tumors
with minimum risk of damage to surrounding healthy tis-
sue. Lasers have also been considered for suturing wounds
through heat. The laser would cauterize the wound as it
sutured it. It could also be used to disinfect small areas
quickly. Dentists have experimented recently with laser
drills; they are fast, sure, and painless.
It is conceivable that laser beams can be made even
narrower in diameter than a single protein molecule. Such
a laser beam might be used to alter the genetic properties
of living organisms.
A team of medical scientists has reported that irradia-
tion by a laser beam has altered the electrical conductiv-
ity of whole human blood. Just what this means or how
it occurs has not yet been made apparent.
SCIENTIFIC APPLICATIONS
Perhaps some of the most far-reaching effects of the
laser will be in the fields of pure and applied science. The
laser may profoundly affect man's understanding of his
natural environment.
Our most basic quantities of measurement are length,
mass, and time. Two of these, length and time, are related
by a constant, the velocity of light in a vacuum, and yet
the value of this constant is only imperfectly known.
Our national standard of frequency is calibrated from
the same astronomical observations that give us our meas-
ure of time, since the frequency of cycles per second that
a wave executes is intimately related to time.
When dealing with radiation in the visible region, scien-
tists measure wavelength instead of frequency. But if the
138
The Laser — What it is and Does
standard radio frequencies could be doubled, redoubled,
and then redoubled again as many times as necessary to
reach the visible-light region, then length and time would
be one and the same thing irrespective of our uncertainty
as to the exact speed of light in a vacuum.
Another basic scientific problem is the question of
whether ether exists or not. You recall that we explained
electromagnetic waves by comparing them to waves in a
pond. Many scientists have found it equally hard to con-
ceive of waves without postulating some substance or
medium in which the waves could move or propagate.
Accordingly, they postulated ether— a colorless, odorless
substance filling all space— in which electromagnetic
waves could propagate just as waves propagate in a pond.
For years now, scientists have been trying to relegate ether
to the same never-never land as phlogiston and other
weird substances once postulated by alchemists to explain
physical phenomena they could not understand.
The first experiment to disprove the existence of ether
was the Michaelson-Morley experiment: If the earth is
rotating in a stationary sea of ether, the ether will drift
by the earth in a direction counter to the earth's rotation.
Now, suppose two light beams are transmitted at right
angles to each other in such a way that the ether drift will
add to the speed of one beam while the other beam will
travel perpendicular to the ether drift and therefore be
unaffected by it. Then any difference in velocity caused by
ether drift could be detected by measuring the difference
in frequency of the two beams. To make the measurement
more precise, the apparatus emitting the light beam is
next turned around so that the ether drift will oppose the
speed of the beam instead of adding to it; the frequency
difference (if any) can again be measured. If the sum
of the two frequency differences were significant, an ether
139
drift could be said to exist. This experiment has been car-
ried out with the use of gas lasers, but no significant
frequency difference has been noticed that could substan-
tiate the existence of an ether.
In the realm of applied science, the laser shows greatest
promise in spectroscopy. We have referred at many times
to absorption of infrared, light, and ultraviolet frequencies
by certain substances. The exact frequencies absorbed de-
pend upon the chemical composition of the substance and
the structure of its molecules. The totality of frequencies
absorbed or the absorption spectrum of a substance is as
individual as your fingerprints. Therefore spectroscopy is
a basic tool for physicists and chemists studying the prop-
erties of matter. But better discrimination in spectroscopy
is needed, and to get it, scientists must know the exact
frequencies with which a substance is irradiated. As the
number of laser materials increases, and consequently the
number of available coherent light frequencies increases,
spectroscopists can look forward to more efficient tools
that will enable them to gain greater and greater insight
into the basic makeup of matter.
Conclusion
In this chapter we have explained the continuum of
the electromagnetic spectrum in terms of both frequency
and wavelength. We have come to grips with some of the
basic concepts of quantum mechanics and have seen how
they explain the action of the three basic types of lasers:
optically pumped, gaseous electrically pumped, and injec-
tion. We have discussed the phenomenon of fluorescence
and have seen how laser action is related to fluorescence
but differs from it because of ( a ) its frequency coherence
or monochromaticity and ( b ) its spatial coherence, or the
fact that all wavelets keep in step.
140
The Laser — What it is and Does
(Incidentally, this last gem of knowledge now makes
you smarter than a certain covey of investors with more
spare cash than technical knowledge. They lost several
kilobucks supporting a glib physicist with a lab full of
bottles of fluorescent material that he passed off as lasers
completely covering the visible spectrum! Of course, they
weren't lasers at all. )
Finally, we have looked at the whole electromagnetic
spectrum in terms of how coherent radiation is or might
be produced by quantum devices, and have placed a
special emphasis on a possible gamma-ray laser. And we
have seen the impact of lasers on national defense, indus-
try, medicine, and science.
Now we shall look backward and see how the laser
actually came into being.
141
One basic law rules the operation of all devices that use
electric currents. A fine introduction to the study of electricity.
8 A Simple Electric Circuit: Ohm's Law
Albert V. Baez
A chapter from the textbook The New College Physics, a Spiral Approach.
WE BEGIN this chapter by considering the
operational steps we might take, in an elementary
laboratory, in order to learn more about electric
current. We shall then try to build up a theory
that accounts for our observations.
40.1. A Simple Series Circuit: Measurement
of Potential Difference
Figure 40.1 shows what our apparatus looks
like: A, a six-volt storage battery; B, a lamp in a
socket; C, a knife switch; D, a voltmeter; £, an
ammeter; F, some connecting wires. From now on
we shall, as much as possible, use the shorthand
of conventional diagrams, as in Figure 40.2, which
shows battery A, lamp B (the zigzag line is actu-
ally the symbol for an element with resistance),
and switch C connected in series. When the switch
is closed, the lamp lights up. We say that there
is an electric current or that there is a flow of
electric charge, but we don't, of course, see any-
thing flowing. The fact that the bulb lights up
when the switch is closed is the only outward
sign that anything flows.
It is not uncommon to begin such an experiment
with little or no knowledge of what is inside the
magic boxes A, B, D, and E (Fig. 40.1). All we
r --— ~
/js;
ft o &
A
■$*'
F
FIG
40.1. Apparatus needed for a simple experiment with electric circuits: A, ^^f^f'j^^uS
C, a knife switch; D, a voltmeter; E, an ammeter; F, typical connecting wires, two oj the clips on winch
are called alligator clips.
143
fig. 40.2. Schematic diagram of a series circuit includ-
ing a battery, A, connected to a lamp, B (shown here
as a resistor), through a switch, C.
know is that D measures potential difference and
that E measures current. In this chapter we shall
look inside B, D, and E. The battery, A, however,
will have to remain just an electron pump; I
shall leave its inner details out of the discussion
because they involve the complicated molecular
mechanism by which chemical energy is converted
into electrical energy.
We want to understand why the voltmeter read-
ings of Figure 40.3 are what they are at different
places. We are going to limit ourselves in this
chapter to an understanding of the simple circuit
of Figure 40.2. We shall move more slowly than
is customary in a chapter on electric circuits, and
only when we peek inside the voltmeter and the
ammeter shall we see slightly more complicated
circuits in series and in parallel. Our immediate
objective is limited ; but, if you understand all the
details of this discussion, you will have a firm
grasp of fundamentals.
We first notice, as we consider the reading of
the voltmeter in different parts of Figure 40.3,
that we do not need to disturb the circuit when we
take a voltmeter reading. We simply connect the
voltmeter to two points of the circuit.
Next we observe and record the data, and then
we try to explain them by theory. When the volt-
meter (Fig. 40.3) is connected across the bat-
tery (A), it reads 6 volts if the switch is open;
with the switch closed (B) it reads 5.45 volts.
Connected across the lamp, it reads 0 if the switch
is open (C) and 5.45 volts if the switch is closed
(D). Connected across one of the connecting wires,
it reads 0 whether the switch is open (E) or
closed (F).
If the voltmeter is telling the truth, the potential
144
difference across the terminals of the battery is
6 volts when there is no current in the circuit (A).
The potential difference across the battery drops
when there is current (B). There is no potential
difference across the terminals of the lamp (C)
until the switch is closed (D), and there is never
a measurable potential difference across one of
the connecting wires. Our theory of what is going
on must account for all these readings (and a lot
more).
Let's begin our description of what we think
is going on. We have already encountered a mo- i
mentary flow of charge in electrostatic experi-
ments, but something different is obviously hap-
pening here, for this current can flow for a long
time. Something replenishes the charge; something '
maintains a potential difference that produces a
steady flow of charge. This something, in our
experiment, is the battery. The terminals of the
battery are charged in the very sense in which we
used the word in electrostatics. If our battery has
only two terminals, an electric field surrounds
6 V (almost)
5.45 V
Zero volts
5.45V
fig. 40.3. Readings on a voltmeter as it is connected to
different parts of a series circuit that is sometimes
open and sometimes closed.
A Simple Electric Circuit: Ohm's Law
fig. 40.4. (A) The electric field lines in air surrounding the terminals of a battery. (B) The electric field lines
within a wire connecting the two terminals of a battery through a lamp.
them as if they constituted an electric dipole.
Figure 40.4.A shows the electric field between the
two battery terminals. It looks very much like the
electric field between two charged metal balls on
insulating stands; but there is a difference in
what happens to these fields if the terminals are
connected with a wire. A wire connecting one
charged ball to the other would carry current
only for an instant, for the potential difference
between them would soon be zero, and the field
would vanish. If the terminals of the battery are
connected, a large current can exist in the wire
for a much longer time, and the field between the
terminals will still be like that of Figure 40.4.A
after the wire is removed. In Figure 40.4. B we
see the electric field lines (E) that come into exist-
ence within the wire that runs from one terminal
through the lamp to the other terminal. I said
earlier that there can be no electric field within a
conductor, but that is true only in the electro-
static case. Charges move in the wire of Figure
40.4.B because there is an electric field within it.
Since the lamp gets hot, it is obvious that energy
is involved. It looks very much as if something
were playing the role that friction plays in me-
chanics. Something is playing that role; it is
called resistance (defined in § 40.3), and we shall
soon consider it in some detail.
Let us now recall the definition of electric field,
E, as F/q, the force per unit charge (§ 4.4). An
electron finding itself in electric field E experi-
ences the force F = — eE. It should experience
the acceleration a = F/m, and it does, but it
cannot pick up much speed, for it collides with
other electrons. The average behavior of many
electrons, starting and stopping, is, nevertheless,
a general drift in the direction of — eE. Statis-
tically, the free electrons drift at an average speed
determined by the magnitude of the force — eE.
The idea of motion at a constant speed under
the action of balanced forces can be perfectly il-
lustrated by the falling of small spheres (such as
marbles) through a tall glass beaker containing
glycerin (Fig. 40.5.A); balls of the right weight
and dimensions achieve a terminal velocity. The
force of gravity, mg, pulls them downward, but a
viscous frictional force, f, pushes them upward.
When wg = f, the acceleration is zero (see § 5.2).
A positive charge, q, in electric field E feels the
force qE (Fig. 40.5. B). If it also feels an equal
retarding force, f, it can move at a constant
speed. What happens in a wire is somewhat like
this. For two reasons, however, you must not
take any such picture literally. First, no electron
travels for long without hitting another, and the
concept of drift velocity is therefore purely statis-
tical. (It takes a lot of kinetic energy to carry an
electron into contact with another, even when the
other is anchored to an atom. What I have called
hitting just means being decelerated by a force
145
fig. 40.5. (A) The gravitational field lines running
through a tall glass beaker containing glycerin; little
spheres fall through it at a constant speed. (B) The
electric field lines in a wire; electric charges move
with a constant average speed within the wire.
field. Here it would pay you to re-read § 3.7,
dealing with the concept of contact.) Second,
electrons have a negative charge and move op-
posite to E, but this does not damage the model
of Figure 40.5.
Traditionally, the direction of current in a wire
has been taken as from the positive to the negative
pole (in the part of the circuit outside the battery).
In this book, since it is now known that in a wire
the electrons do the moving, I have broken with
tradition by assigning to / the direction of elec-
tron flow. But I shall use the symbol /(= -/) for
the conventional direction (from positive to nega-
tive) whenever it can simplify the wording of
statements. All the left-hand rules I gave in the
study of magnetism relate to /. If we associate
the right hand with /, similar rules apply. In
other words, / is the direction in which positive
charges would move in a wire. Since positive
charges tend to move from a region of high elec-
tric potential to one of low potential, it is con-
venient to use the traditional symbol for current, j
/, in these cases. (We simply need to remember
that the electrons in metallic conductors move in
the opposite direction; in liquids, however, posi-
tive as well as negative charged bodies move.)
Whenever we use the symbol q without any further
specification, it will represent a positive charge.
The electronic charge will, of course, be written as
— e.
There are two ways of expressing the reason
why a ball moves downward through the beaker
of glycerin. One is to say that it moves down
because wg points downward; the other is to say
that it has a tendency to move from a region of
high potential to one of low potential. The same
language applies to positive charges in an electric
field: they move from A to B in Figure 40. 5. B
because qE points that way, or (since an applied
force would do work in moving a positive charge
from B to A) they move from a region of high
potential to one of low potential.
Potential difference, V, is measured in volts,
which we identified earlier (§ 37.2) with joules per
coulomb. The work that will move charge Iq
through distance x from B to A is (by the formula
"work equals force times distance") MJ = (Aq)Ex.
The work per unit charge is AU Aq = Ex. The left-
hand side has the units joules per coulomb, or i
volts. The right-hand side has newtons per cou-
lomb times meters for units. This equivalence is
worth remembering. We may write
or
V = Ex
E = V/x
[40.1
[40.2
Now we are getting somewhere. The quantities
on the right-hand side of the second equation
are measurable, V with a voltmeter (we'd better
find out how it works), x with a meter stick.
If we connected a voltmeter across points A
and B of Figure 40.5. B, would it show a reading?
I said earlier (Fig. 40.3. F) that there is no de-
tectable reading across a wire carrying current.
You will have to take my word for it that a cer-
tain very sensitive kind of voltmeter would indi-
cate a small potential difference between points
A and B if there were a current in the wire.
146
A Simple Electric Circuit: Ohm's Law
EXAMPLE 40.1. A sensitive voltmeter indicates a
potential difference of 10-6 V between points A and
B of Fig. 40.5. B. The distance between the points
is x = 2 m. We wish to know (1) what force, in
newtons, an electron feels within the wire; (2) what
acceleration it experiences; (3) what the increment
in its speed is if it travels for 10-7 sec.
1. The force on a charge, q, is F = Eq. Since, by
equation 40.2, E = V/x, we know that F = Vq/x.
We are given that
V = 10~6 V
q = -e = -1.60Xl0-19coul
x = 2m
Therefore, if we drop the minus sign,
10~6 X 1.60X10"19
F =
nt
= 8Xl0-26nt
2. The acceleration is a = F/m. We know that
F = 8xl0"26nt
m = 9.11Xl0-31kg
Therefore
8Xl0~26nt
" 9.11XlO"31kg
= 8.78xl04m/sec2
3. We know that
Av/At = a
Therefore
Av = a(At)
- 8.78 XlO4 m/sec2 X lO"7 sec
= 8.78X10-3 m/sec
There are experimental reasons for believing that
this is of the right order of magnitude for the aver-
age speed of electrons in a wire.
40.2. Electromotive Force
We can extend the analogy of balls falling
through glycerin to a simple electric circuit.
In Figure 40. 6. A we see balls rolling and falling
under the action of the earth's gravitational field,
g. If the balls are to keep moving at a constant
rate, work has to be done against gravitational
force as each ball is lifted from D to A. The energy
is supplied by the man, who acquires it by the
complicated chemical process that transforms food
energy into mechanical energy. Notice that there
is a small difference in gravitational potential,
g(Ahi), between points A and B, a large differ-
Ah,
Ah,
^
I Ah,
— <
B
fig. 40.6. Analogy between the effect of the earth's gravitational field and that of an electric field.
147
■*"" — 5"*"
i.
2. 73 amp
2. 73 amp
FIG. 40.7. How an ammeter will read when connected
in different parts of a circuit.
ence, g(Ah2), between points B and C, and a small
difference again, g(Ah3), between points C and D.
In this arrangement a "potential-difference meter"
(analogous to a voltmeter) could consist of an
ordinary meter stick.
The frictional force on each ball as it falls in
the glycerin from B to C is equal to its weight.
This makes the resultant force zero, which is
what is required for descent at a constant speed.
The frictional force on each ball in AB and CD
is much smaller than its weight. This is suggested
by the small slope of the inclined planes in these
regions. The man has to do work mgh (/; =
A/?i + A/z2 + A/73) on each ball to move it from
D back to A so that it can start the cycle again.
In Figure 40.6. B we have the electrical counter-
part of Figure 40.6. A, a complete electric circuit,
ABCD. Electric charges are moving under the
influence of the electric field, E. The potential
difference between points A and B is very small
because the charges encounter only a slight re-
sistance to their motion in this region. The poten-
tial difference between points B and C is great
because the resistance there is great; the letter R
signifies, in fact, that this portion of the circuit,
like the lamp in Figure 40.2, is a resistor (a con-
ductor with relatively large resistance). There is
only a small potential difference between C and D.
The charges have a low potential at D, and it
takes energy, which is supplied by the battery,
to lift them to a high potential at A. The battery
transforms chemical into electrical energy by a
complicated process, which I shall not analyze
any more than I analyzed the internal workings
of the man of Figure 40. 6. A.
The ability of the battery to raise positive
charges from a low potential at D to a high po-
tential at A is measured by the number of joules
per coulomb, AW/Aq, it needs in order to do this.
(It is actually electrons, with negative charges,
that are moving — and the other way round; but
this poses only semantic problems. We could talk
the whole thing out by using different words, but
we are here adhering to the classical idea that
current consists of positive charges whose poten-
tial is raised in going from D to A.) The ratic
AW/Aq is called the electromotive force (abbrevi-
ated as emf) of the battery and is symbolizec
as 8. It is the work per unit charge done by the
battery in moving positive charges against the
electric field within the battery. It is not, ot
course, a force in the Newtonian sense; it i<
measured in joules per coulomb, or volts, not ir
newtons; but the word "force" has become firmh
established in the vocabulary of electricity. Since
AW/Aq is measured in volts, you might ask wh\
we do not simply say that 8 is the difference ir
potential between points D and A. The answei
is that the battery itself may have internal re
sistance, and that the potential difference betweer
points D and A may therefore be somewhat les:
than 8, depending on how much internal re
sistance there is. Ideally, with no internal resist
ance, 8, measured in volts, would be equal t(
the difference in potential between points D and A
Let us return, for illustration, to Figure 40.3
The voltmeter showed (B) a potential different
of 5.45 volts between D and A when there wa.
electric current in the circuit. This was not, how
ever, the emf of the battery. The potential differ
ence across the terminals of a battery is neve
exactly equal to its emf when there is curren
through the battery. When the switch is open (A)
the potential difference is almost 6 volts. We hav
to hedge here because some charges flow evei
when the voltmeter alone is connected across th
battery; the potential difference is not quite equa
to the emf unless the resistance of the voltmete
is infinite — that is, unless the voltmeter draw> n
current. A good voltmeter, obviously, has a ver
high resistance.
148
A Simple Electric Circuit: Ohm's Law
I have been using the term "resistance" in a
qualitative way. In order to define it precisely, I
have to measure current. Notice that the argu-
ment so far has not depended upon current. I
have talked only of potential difference ("volt-
age" in the vernacular of the electrician). But
perhaps our rolling-ball analogy (Fig. 40.6) has
shown why the reading of the voltmeter in Fig-
ure 40.3. F was zero. (It corresponded to a van-
ishingly small Ahi.) The potential rise (8 = ArV/Aq)
within the battery — that is, the emf — must equal
the sum of the potential drops (AV) in the complete
circuit or loop. We let VAB mean "the potential
difference between points A and 5." Since VAD
and VCd (Fig. 40.6. B) are both practically zero,
the voltmeter readings of Figure 40.3.B,D are
practically identical. We now imagine (Fig. 40.6)
connecting one terminal of the voltmeter to point
A. We then touch points B, C, and D with a wire
connected to the other terminal of the voltmeter.
We read that VAB = 0, that VAC = 5.45 volts,
and that VAD = 5.45 volts. The reason for this is
that
VAD = VAB + VBC + VCD
= 0 + 5.45 V + 0 = 5.45 V
Before we can proceed, we need to define resist-
ance in terms of potential difference and current.
40.3. Ohm's Law
We shall now use the ammeter in the circuit of
Figure 40.3. To use an ammeter, you must break
into the circuit at some point and allow the current
to pass through the ammeter.^ Figure 40.7 shows
that the ammeter reads 2.73 amperes in each of
four different positions. This simply means that
charges are conserved. The number of charges
flowing per second past any point in the circuit
must be the same as the number flowing per sec-
ond past any other point; otherwise charges would
be either accumulating or leaking away. If, for
t Two interesting exceptions to this statement are: (1) a
special alternating-current ammeter that just clamps its
coil round the current-carrying wire; (2) a special direct-
current meter, used by automobile electricians, that works
essentially like Oersted's experiment; it is simply clipped
onto the battery-charging line.
fig. 40.8. The sum of inward currents at a junction is
equal to the sum of outward currents.
fig. 40.9. One way to connect an ammeter and a volt-
meter to measure the resistance of a resistor.
example (Fig. 40.8), we have a junction, O, where
the currents are T, /->, h, and /4, it must be true
that 2/ = 0— that is, that h + 72 + h + h = 0—
if we consider "coming into O" as positive and
"leaving 0" as negative.
So far Figure 40.7 simply confirms the fact that
the current in a single loop is the same every-
where, including the battery. Outside the battery,
positive charges tend to flow from regions of high
to regions of low potential; inside the battery,
the energy supplied by the battery makes it pos-
sible for positive charges to flow against the
electric field that is naturally there (compare DA
in the rolling-ball analogy, Figure 40.6. A).
We now need an experimental fact about metal-
lic conductors. If such a conductor (labeled BC)
is connected as in Figure 40.9, the ammeter will
show the current in it, and the voltmeter will
show the voltage across it. If different currents, I,
are made to flow through it, different voltages,
V, will appear across it. A plot of V agaiust / is
a straight line going through the origin (Fig.
40.10); that is, the ratio of V to I is constant.
(This is not true of all kinds of conductors; it is
149
not true, for example, of vacuum tubes or of
certain types of crystals.) I shall now define, by
the following equation, the quantity called the
resistance, R, of the conductor BC:
R =
[40.3
For some materials (for many different kinds
of metallic wires, for example) and under certain
conditions (at constant temperature, for exam-
ple) the resistance defined in this way is a con-
stant, independent of /. For other kinds of
conductors (vacuum tubes, for example) the R
defined in this way is not independent of /. In
all cases the resistance defined by equation 40.3
is measured in ohms. Obviously, "volts divided
by amperes" is equivalent to ohms. Equation 40.3
is known as Ohm's law after Georg Simon Ohm,
a German physicist (1787-1854).
If the current is / and the cross-sectional area
■MMMiinim maw— ummmw i asMmmmnsi
fig. 40.10. A plot of voltage against current in an olvnie
conductor.
— — 7^\%*i
fig. 40.1 1. Illustrating the definition of current density.
150
fig. 40.12. The voltage drop between Vv and P: is sc
small that the bird feels no shock.
of the wire is A, the current density, j, has the
magnitude
/
J =
[40.4-
and is measured in amperes per square meter.
For the class of conductors I have been talking
about (called ohmic conductors) it is an experi-
mental fact that the electric field intensity, E^
established inside the wire (Fig. 40.1 1) is prop
tional to the current density in the wire. In ot
words, experiments show that
por-
Eoc j
[40 J
(I have written j as a vector because E is a vector.)
There must be a constant of proportionality, p,.
such that
E = pj [40.6\
Remembering that E is measured in volts per
meter (equation 40.2), let us find the potential
difference, V, across a length, /, of wire as follows.
Dropping the vector notation, we have
El = pjl [40./ 1
Using equation 40.4, we get
/
£/ = „-/
[40.6:
But, according to equation 40.2, El = V. There-
fore
K=p-/
[40.9
A Simple Electric Circuit: Ohm's Law
.^
fig. 40.13. The voltage drop between Qi and Q2 might be great enough to kill the bird.
or
Pi
A
[40.10
But this is the ratio that defines resistance, R
; (equation 40.3). Hence
R-pl
R~A
[40.11
That is, the resistance of a wire is directly propor-
tional to its length and inversely proportional to
its cross-sectional area. [I could have introduced
p by means of equation 40.11, but I wanted to
emphasize, once again (equation 40.6), the exist-
ence of an electric field within a wire carrying a
current.] The constant of proportionality, p, is
' called the resistivity of the material. Resistivity is
the inverse of conductivity. Table 37.2 lists the
; resistivities of some common substances.
EXAMPLE 40.2. We wish to find the resistance of a
piece of copper wire 1 km long and 1 mm in di-
ameter.
We know that
p = 0.172 X10-7 ohm-meter
/ = 103m
d = 10-3 m
Therefore
A = ?$- = 7.85 X 10-7 m2
4
and (equation 40.11)
0. 1 72 Xl0~7 ohm-meter X 103m
R =
7.85X10"
= 21.9 ohms
m-
(The filament of an ordinary 100-W light bulb
has a resistance of about 100 ohms.)
We can now consider the voltage drop in wires
carrying current. You have seen birds perched on
such wires without being killed and apparently
without feeling any shock. Now, one of the harm-
ful things in electric shock, to birds or to people,
is the current through the body. This current
obeys, approximately, Ohm's law, which implies
that we get big currents through the body if we
touch points with large potential differences.
There is a voltage drop, V = IR (see equation
40.3), in a wire, but the potential difference (Fig.
40.12) between points Pi and P2, where the bird's
feet rest on the wire, is exceedingly small. In
Example 40.2 we saw that the resistance of 1,000
meters of a certain copper wire was 21.9 ohms.
The resistance of 10 centimeters would be only
21.9X 10~4 ohm. Even if the wire carried a current
of 100 amperes (very unlikely), the potential drop
from Pi to P2 would be only 0.219 volt. Such a
small potential difference could not send enough
current through the bird to do much harm.
A great potential drop might occur (Fig. 40.13)
across some distant load — a motor, M, perhaps.
151
Hence the potential difference between points Qx
and Qi on wires carrying the same current might
be very great indeed. If the bird could put one
foot at Q\ and the other at Q2, it might be killed.
We can now consider our original circuit sym-
bolically. In Figure 40.14 the battery, B, with its
internal resistance, r, is enclosed in a dashed
line; the lamp, L, has resistance R. The current,
/, is the same in both B and L. The potential
drop in L is IR (equation 40.3); the potential drop
in B is Ir. The charges leave point P at the same
potential at which they arrive there. The work per
unit charge done by the battery, 8 = AW/Aq,
must therefore exactly equal the drop in potential,
IR + Ir. Hence
A
Ir = IR
[40.12
Now Figure 40.3 indicates (D) that IR = 5.45
volts and (B) that 8 — Ir = 5.45 volts. From
Figure 40.7 we see that / = 2.73 amperes. There-
fore
5 45 V
R = z^ = 2 ohms [40.13
2.73 amp L
From Figure 40.3. A we know that S is almost
6 volts. Therefore, using the equation
8 - Ir - 5.45 V [40.14
we get
Ir = (6 - 5.45) V
- 0.55 V [40.15
But / = 2.73 amperes. Therefore
= 0.55 V
2.73 amp
= 0.2 ohm
We have now accounted for the voltage readings
of Figure 40.3, and we have learned something
about electric circuits in the process.
40.4. How the Ammeter and the Voltmeter
Work
I have already told how a galvanometer works;
it is a coil, mounted between the poles of a mag-
net, whose dipole moment experiences a torque
when it carries current (§ 38.4). If (Fig. 40. 15.A,C)
a low-resistance conductor, S (called a shunt), is
connected across the coil, C, in parallel with it,
most of the current flows through S, and we have
fig. 40.14. Our original series circuit treated symboli-
cally. The internal resistance of the battery is shown
as r. // 8 is the emf of the battery, 8 — Ir = IR.
an ammeter. The combination, which has a low
resistance, can be designed to measure even a
large current, for very little of the current flows
through the coil.
The same galvanometer can be converted into
a voltmeter (Fig. 40.15.B,D). If the coil, C, is
connected in series with a resistor, M, of high
resistance (called a multiplier), even a large poten-
tial difference, V, across the terminals A and B
will produce only a small current through the
coil, C, since / = V/R and R here includes the
resistance of both M and C. The whole device
has, as a good voltmeter must have, a high resist-
ance. The details may be clarified by reference to
Problems 40.15, 40.16, and 40.17.
It is also left for Problem 40.14 to prove that,
when two resistors are connected in series, the
resistance of the combination is simply the sum
of the two resistances, but that, when they are
connected in parallel, the reciprocal of the combi-
nation is the sum of the reciprocals of the indi-
vidual resistors. For resistors in series (as in
Fig. 40.15.B.D)
R = Ri + /?- [40.16
For resistors in parallel (as in Fig. 40. 15. A, C)
R /?, R2
[40.17
40.5. Electric Power Dissipated as Heat
The analogy of balls falling through glycerin
(Fig. 40.6.A) is useful, for we see immediately
that the loss in potential energy must appear as
heat. Similarly, the loss in potential energy of
charges moving in the resistor, R, of Figure 40.6.B
can appear in the form of heat. The work re-
152
A Simple Electric Circuit: Ohm's Law
quired to lift a bail in Figure 40.6. A is W = mgh.
The work per unit mass is W/m = gh. Similarly,
the work required to move the positive charge
Aq from B to A is
AW = (Aq)VBA = coulombs X volts
joules
coulombs X
joules
[40.18
coulomb
The rate of doing work, P (for power), is
At At BA
But Aq/At is the current, /, in amperes. Hence
P = IV B a- This must be in joules per second, or
watts. If all this power goes into heating the
resistor, we may write
Pj = IV [40.19
The subscript J is for "joule," to remind us that
heat is being generated. Thus "amperes times
volts" is equivalent to "watts." Since 4.184
joules = 1 calorie, we may use the expression
IV/4AS4 to compute the calories per second gen-
erated in a resistor.
From Ohm's law (equation 40.3) we know that
V = IR ; so we may write
Pj = i(iR) = pr [40.20
Since, if several resistors are connected in series,
they all carry the same /, this form of the equation
(Pj = PR) is useful.
On the other hand, since / = V/R, we may
write
V V2
P< = -V = —
J R R
[40.21
A9
S
A/WWV
■ — >■
92?
1ZZ
A 9
92?
! nfooo<n i
Ammeter
FIG. 40.15. Symbolic representation of the components (A, C) of an ammeter and (B, D) of a voltmeter.
153
Since, if several resistors are connected in parallel,
each has the same potential drop as the others,
this form (Pj = V2/R) is applicable to such com-
binations.
40.6. Summary
A battery has the ability to raise positive charges
from a low potential to a high potential. Positive
charges in an external electric circuit connected
to this battery tend to flow from the region of
high potential to that of low potential. This flow
is called current. Actually, in wires, negative
charges (electrons) flow in the opposite direction,
but the logic is not affected.
The work per unit charge done by the battery
is called its electromotive force, 8; it is the ratio
AW/Aq, measured in joules per coulomb, or volts.
The potential rise in the battery must equal the
sum of all the potential drops, AV, in the whole
circuit. The potential drop across an ohmic re-
sistor of resistance R in which there is current /
is V — IR (Ohm's law). The resistance of a wire is
directly proportional to the product of its length
and its resistivity and is inversely proportional to
its cross-sectional area.
The flow of charges in a wire is very similar to
the flow of a liquid in a pipe. When several wires
meet at a point, for example, the sum of the in-
ward currents is equal to the sum of the outward
currents.
In both pipes and wires energy can be dissipated
in the form of heat. If the potential drop in a wire
is V, the work it takes to move charge q across
it is qV, and the rate of doing work, or power, is
P = IV. The power that goes into heating a
resistor may be written as PR or as V2/R.
154
A brief, informal review of the electronic age, past
and present.
The Electronic Revolution
Arthur C. Clarke
An excerpt from his book, Voices from the Sky, originally published
in the New York Times in 1962.
The electron is the smallest thing in the universe; it would
take thirty thousand million, million, million, million of them
to make a single ounce. Yet this utterly invisible, all but
weightless object has given us powers over nature of which
our ancestors never dreamed. The electron is our most ubiqui-
tous slave; without its aid, our civilization would collapse in a
moment, and humanity would revert to scattered bands ol
starving, isolated savages.
We started to use the electron fifty years before we dis-
covered it. The first practical application of electricity (which
is nothing more than the ordered movement of electrons)
began with the introduction of the telegraph in the 1840's.
With really astonishing speed, a copper cobweb of wires and
cables spread across the face of the world, and the abolition of
distance had begun. For over a century we have taken the
instantaneous transfer of news completely for granted; it is
very hard to believe that when Lincoln was born, communi-
cations were little faster than in the days of Julius Caesar.
Although the beginning of "electronics" is usually dated
around the 1920^, this represents a myopic view of tech-
nology. With the hindsight of historical perspective, we can
now see that the telegraph and the telephone are the first two
landmarks of the electronic age. After Alexander Graham Bell
had sent his voice from one room to another in 1876, society
could never be the same again. For the telephone was the first
155
electronic device to enter the home and to affect directly the
lives of ordinary men and women, giving them the almost
godlike power of projecting their personalities and thoughts
from point to point with the speed of lightning.
Until the closing years of the nineteenth century, men used
and handled electricity without knowing what it was, but in
the 1890's they began to investigate its fundamental nature,
by observing what happened when an electric current was
passed through gases at very low pressures. One of the first,
and most dramatic, results of this work was the invention of
the X-ray tube, which may be regarded as the ancestor of all
the millions of vacuum tubes which followed it. A cynic
might also argue that it is the only electronic device wholly
beneficial to mankind— though when it was invented many
terrified spinsters, misunderstanding its powers, denounced
poor Rontgen as a violator of privacy.
There is an important lesson to be learned from the X-ray
tube. If a scientist of the late Victorian era had been asked
"In what way could money best be spent to further the
progress of medicine?" he would never by any stretch of the
imagination have replied: "By encouraging research on the
conduction of electricity through rarefied gases." Yet that is
what would have been the right answer, for until the dis-
covery of X rays doctors and surgeons were like blind men,
groping in the dark. One can never predict the outcome of
fundamental scientific research, or guess what remote and
unexpected fields of knowledge it will illuminate.
X rays were discovered in 1895— the electron itself just one
year later. It was then realized that an electric current consists
of myriads of these submicroscopic particles, each carrying a
minute negative charge. When a current flows through a solid
conductor such as a piece of copper wire, we may imagine the
electrons creeping like grains of sand through the interstices
between the (relatively) boulder-sized copper atoms. Any
individual electron does not move very far, or very fast, but it
jostles its neighbor and so the impulse travels down the line at
156
The Electronic Revolution
speeds of thousands of miles a second. Thus when we switch
on a light, or send a Morse dash across a transatlantic cable,
the response at the other end is virtually instantaneous.
But electrons can also travel without wires to guide them,
when they shoot across the empty space of a vacuum tube like
a hail of machine-gun bullets. Under these conditions, no
longer entangled in solid matter, they are very sensitive to the
pull and tug of electric fields, and as a result can be used to
amplify faint signals. You demonstrate the principle involved
every time you hold a hose-pipe in your hand; the slightest
movement of your wrist produces a much greater effect at the
far end of the jet. Something rather similar happens to the
beam of electrons crossing the space in a vacuum tube; they
can thus multiply a millionfold the feeble impulses picked up
by a radio antenna, or paint a fluorescent picture on the end
of a television screen.
Until 1948, electronics was almost synonymous with the
vacuum tube. The entire development of radio, talkies, radar,
television, long-distance telephony, up to that date depended
upon little glass bottles containing intricate structures of wire
and mica. By the late logo's the vacuum tube had shrunk
from an object as large as (and sometimes almost as luminous
as) an electric light bulb, to a cylinder not much bigger than a
man's thumb. Then three scientists at the Bell Telephone
Laboratories invented the transistor and we moved from the
Paleoelectronic to the Neoelectronic Age.
Though the transistor is so small-its heart is a piece of
crystal about the size of a rice grain-it does everything that a
radio tube can do. However, it requires only a fraction of the
power and space, and is potentially much more reliable. In-
deed, it is hard to see how a properly designed transistor can
ever wear out; think of little Vanguard I, still beeping away
up there in space, and liable to continue indefinitely until
some exasperated astronaut scoops it up with a butterfly net.
The transistor is of such overwhelming importance because
it (and its still smaller successors) makes practical hundreds
157
of electronic devices which were previously too bulky, too ex-
pensive or too unreliable for everyday use. The pocket radio is
a notorious example; whether we like it or not, it points the
way inevitably to a day when person-to-person communica-
tion is universal. Then everyone in the world will have his
individual telephone number, perhaps given to him at birth
and serving all the other needs of an increasingly complex
society (driving license, social security, credit card, permit to
have additional children, etc. ) . You may not know where on
Earth your friend Joe Smith may be at any particular mo-
ment; but you will be able to dial him instantly— if only you
can remember whether his number is 8296765043 or
8296756043.
Obviously, there are both advantages and disadvantages in
such a "personalized" communication system; the solitude
which we all need at some time in our lives will join the
vanished silences of the pre-jet age. Against this, there is no
other way in which a really well-informed and fast-reacting
democratic society can be achieved on the original Greek
plan— with direct participation of every citizen in the affairs
of the state. The organization of such a society, with feedback
in both directions from the humblest citizen to the President
of the World, is a fascinating exercise in political planning. As
usual, it is an exercise that will not be completed by the time
we need the answers.
A really efficient and universal communications system,
giving high-quality reception on all bands between all points
on the Earth, can be achieved only with the aid of satellites.
As they come into general use, providing enormous informa-
tion-handling capacity on a global basis, today's patterns of
business, education, entertainment, international affairs will
change out of all recognition. Men will be able to meet face
to face (individually, or in groups) without ever leaving their
homes, by means of closed circuit television. As a result of
this, the enormous amount of commuting and traveling that
now takes place from home to office, from ministry to United
158
The Electronic Revolution
Nations, from university to conference hall will steadily de-
crease. There are administrators, scientists and businessmen
today who spend about a third of their working lives either
traveling or preparing to travel. Much of this is stimulating,
but most of it is unnecessary and exhausting.
The improvement of communications will also render obso-
lete the city's historic role as a meeting place for minds and a
center of social intercourse. This is just as well anyway, since
within another generation most of our cities will be strangled
to death by their own traffic.
But though electronics will ultimately separate men from
their jobs, so that (thanks to remote manipulation devices)
not even a brain surgeon need be within five thousand miles
of his patient, it must also be recognized that few of today's
jobs will survive long into the electronic age. It is now a cliche
that we are entering the Second Industrial Revolution, which
involves the mechanization not of energy, but of thought.
Like all cliches this is so true that we seldom stop to analyze
what it means.
It means nothing less than this: There are no routine, non-
creative activities of the human mind which cannot be carried
out by suitably designed machines. The development of com-
puters to supervise industrial processes, commercial transac-
tions and even military operations has demonstrated this
beyond doubt. Yet today's computers are morons compared to
those that they themselves are now helping to design.
I would not care to predict how many of today's professions
will survive a hundred years from now. What happened to the
buggywhip makers, the crossing sweepers, the scriveners, the
stonebreakers of yesteryear? (I mention the last because I can
just remember them, hammering away at piles of rock in the
country lanes of my childhood. ) Most of our present occupa-
tions will follow these into oblivion, as the transistor inherits
the earth.
For as computers become smaller, cheaper and more re-
liable they will move into every field of human activity. Today
159
they are in the office; tomorrow they will be in the home.
Indeed, some very simple-minded computers already do our
household chores; the device that programs a washing ma-
chine to perform a certain sequence of operations is a special-
ized mechanical brain. Less specialized ones would be able to
carry out almost all the routine operations in a suitably de-
signed house.
Because we have so many more pressing problems on our
hands, only the science-fiction writers— those trail-blazers of
the future— have given much thought to the social life of the
later electronic age. How will our descendants be educated for
leisure, when the working week is only a few hours? We have
already seen, on a worldwide scale, the cancerous growths
resulting from idleness and lack of usable skills. At every
street corner in a great city you will find lounging groups of
leather-jacketed, general-purpose bioelectric computers of a
performance it will take us centuries and trillions of dollars to
match. What is their future— and ours?
More than half a century ago H. G. Wells described, in
The Time Machine, a world of decadent pleasure lovers,
bereft of goals and ambitions, sustained by subterranean ma-
chines. He set his fantasy eight hundred thousand years in the
future, but we may reach a similar state of affairs within a
dozen generations. No one who contemplates the rising curve
of technology from the Pilgrim fathers to the Apollo Project
dare deny that this is not merely possible, but probable.
For most of history, men have been producers; in a very few
centuries, they will have to switch to the role of consumers,
devoting their energies 100 per cent to absorbing the astro-
nomical output of the automated mines, farms and factories.
Does this really matter, since only a tiny fraction of the
human race has ever contributed to artistic creation, scientific
discovery or philosophical thought, which in the long run are
the only significant activities of mankind? Archimedes and
Aristotle, one cannot help thinking, would still have left their
marks on history even if they had lived in a society based on
160
The Electronic Revolution
robots instead of human slaves. In any culture, they would be
consumers of goods, but producers of thought.
We should not take too much comfort from this. The elec-
tronic computers of today are like the subhuman primates of
ten million years ago, who could have given any visiting
Martians only the faintest hints of their potentialities, which
included the above mentioned Archimedes and Aristotle.
Evolution is swifter now; electronic intelligence is only dec-
ades, not millions of years, ahead.
And that— not transistor radios, automatic homes, global
TV— is the ultimate goal of the Electronic Revolution.
Whether we like it or not, we are on a road where there is no
turning back; and waiting at its end are our successors.
161
No. 223,898.
T. A. EDISON.
Electric-Lamp.
Patented Jan. 27, 1880.
0
A
k
€%
h4
i::!'
dnwvrUot
3llC/fU^ (I. SdtdCTV
/" ^^^Tfro^i^O
<«'J.
EDISON'S PATENT on the incandescent lamp was accompanied thickened ends of filament (c), platinum wires Id), clamp, In),
by this drawing. The labeled parte are the carbon filament (o), leading wire, (x), copper wire, (e), tube to vacuum pump (m).
162
Thomas Edison's incandescent lamp, generally as-
sumed to be the product of inspired tinkering, was
actually only one element in a more far-reaching in-
vention: an entire system of lighting.
The Invention of the Electric Light
Matthew Josephson
An article trom Scientific American, 1959.
I
can hire mathematicians, but
mathematicians can't hire me!"
By such declarations in the time
of his success and world-wide fame
Thomas Alva Edison helped to paint his
own portrait as an authentic American
folk hero: the unlettered tinkerer and
trial-and-error inventor who achieved
his results by persistence and a na-
tive knack for things. He is said, for ex-
ample, to have tried more than 1,600
lcinds of material ("paper and cloth,
thread, fishline, fiber, celluloid, box-
wood, coconut-shells, spruce, hickory,
hay, maple shavings, rosewood, punk,
cork, flax, bamboo and the hair out of a
red-headed Scotchman's beard") until
he hit upon the loop of carbonized cot-
ton thread that glowed in a vacuum for
more than half a day on October 21,
1879. Today, in a world that relies for
its artificial illumination largely on his
incandescent lamp, this invention is not
regarded as an especially profound con-
tribution to technology. It rates rather as
a lucky contrivance of Edison's cut-and-
try methods— of a piece with his stock
ticker, mimeograph machine, phono-
graph and alkaline storage-battery— in
the esteem of a public that has come to
appreciate the enormous practical signif-
icance of higher mathematics and ab-
struse physical theory.
If Edison's contribution to the light
of the world consisted solely in the se-
lection of a filament, this estimate of his
person and achievements might be al-
lowed to stand. But the history that is
so obscured by legend tells quite an-
other story. Edison's electric light was
not merely a lamp but a system of elec-
tric lighting. His invention was an idea
rather than a thing. It involved not only
technology but also sociology and eco-
nomics. Edison was indisputably the
first to recognize that electric lighting
would require that electricity be gen-
erated and distributed at high voltage
in order to subdivide it among a great
many high-resistance "burners," each
converting current at low amperage
(that is, in small volume) with great
efficiency into light.
In the 15 months between the time
he conceived his invention and the date
on which he demonstrated it to the pub-
lic, Edison and his associates designed
and built a new type of electric genera-
tor, successfully adapted the then much-
scorned parallel or "multiple-arc" circuit
that would permit individual lights to
be turned on or off separately and, last
of all, fashioned a lamp to meet the
specifications of his system. The labora-
tory notebooks of those months of fran-
tic labor show the Wizard of Menlo Park
endowed with all the prodigious capaci-
ties attributed to him by contemporary
legend. They show in addition that this
self-taught technologist was possessed of
a profound grasp of the nature of elec-
tricity and an intuitive command of its
logic and power.
It was on September 8, 1878, that
Edison was inspired to devote his talents
full time to the challenge of electric
lighting. On that day he went to An-
sonia, Conn., to visit the brass-manufac-
turing plant of William Wallace, co-
inventor with Moses G. Farmer of the
first practical electric dynamo in the
U. S. Wallace showed Edison eight bril-
liant carbon-arc lights of 500 candle-
power each, powered by a dynamo of
eight horsepower. It was with such a
system that Wallace and Farmer, as well
as Charles Brush of Cleveland, were
then beginning to introduce the electric
light on a commercial scale, for street-
lighting and for illuminating factories
and shops. Farmer had made the first
demonstration of arc-lighting in this
country two years earlier, at the Cen-
tennial Exposition in Philadelphia, and
John Wanamaker's store in that city was
already illuminated with arc lights.
Carbon arcs are still employed in
searchlights and in theater floodlights
and projectors to produce light of high
intensity. The current crossing a small
gap between the electrodes creates an
arc. Ionization and oxidation of the car-
bon in the heat of the arc generate a
brilliant blue-white light.
In the 1870's Europe was a decade
ahead of the U. S. in the technology of
arc-lighting. Stores, railway stations,
streets and lighthouses in Britain and
France were equipped with arc lights.
Shedding an almost blinding glare, they
bumed in open globes that emitted nox-
ious gases, and they could be employed
only high overhead on streets or in pub-
lic buildings. Since they consumed large
amounts of current, they had to be wired
in series, that is, connected one to an-
other in a single continuous circuit so
that all had to be turned on or off to-
gether. The multiple-arc circuit, with the
lights connected as in the rungs of a lad-
der between the main leads of the cir-
cuit, was not adapted to such systems
and was considered prohibitive in cost.
Edison himself had experimented
with arc lights, using carbon strips as
burners. He had also investigated the
Editor's Note
The author has based this article
on material in his biography Edi-
son, just published by the McGraw-
Hill Book Company. Copyright ©
1959 by Matthew Josephson.
163
incandescent light, as had many inven-
tors before him. But the slender rod or
pencil of carbon or metal would always
burn up, sooner rather than later, upon
being heated to incandescence by the
current. It would do so though substan-
tially all of the air had been pumped out
of the glass envelope in which it was
contained. Edison had abandoned the
effort to devote himself to a more prom-
ising invention: the phonograph.
Now at Wallace's establishment, con-
fronted with the achievements of others
in the field, he regained his earlier en-
thusiasm. As an eyewitness recalled,
"Edison was enraptured. ... He fairly
gloated. He ran from the instruments
[the dynamos] to the lights, and then
again from the lights back to the electric
instruments. He sprawled over a table
and made all sorts of calculations. He cal-
culated the power of the instruments
and the lights, the probable loss of power
in transmission, the amount of coal the
instrument would use in a day, a week,
a month, a year."
To William Wallace he said challeng-
ingly: "I believe I can beat you making
the electric light. I do not think you are
working in the right direction." They
shook hands in friendly fashion, and
with a diamond-pointed stylus Edison
signed his name and the date on a goblet
provided by his host at dinner.
From Edison's own complete and ex-
plicit notebooks and from the buoyant
interviews that he gave to the press at
this time we know what made him feel
in such fine fettle as he left Wallace's
plant. "I saw for the first time everything
in practical operation," he said. "I saw
the thing had not gone so far but that I
had a chance. The intense light had not
been subdivided so that it could be
brought into private houses. In all elec-
tric lights theretofore obtained the light
was very great, and the quantity [of
lights] very low. I came home and made
experiments two nights in succession. I
discovered the necessary secret, so sim-
ple that a bootblack might understand
it. . . . The subdivision of light is all
right."
The Subdivision of Light
At this time there flashed into Edison's
mind the image of the urban gas-lighting
system, with its central gashouse and gas
mains running to smaller branch pipes
and leading into many dwelling places at
last to gas jets that could be turned on
or off at will. During the past half-cen-
tury gas-lighting had reached the stature
of a major industry in the U. S. It was
restricted, of course, to the cities; three
fourths of the U. S. population still lived
in rural areas by the dim glow of kero-
sene lamps or candles. Ruminating in
solitude, Edison sought to give a clear
statement to his objective. In his note-
book, under the title "Electricity versus
Cas as a General Illuminant," he wrote:
"Object: E. ... to effect exact imita-
tion of all done by gas, to replace light-
ing by gas by lighting by electricity. To
improve the illumination to such an ex-
tent as to meet all requirements of nat-
ural, artificial and commercial condi-
tions. . . . Edison's great effort— not to
make a large light or a blinding light,
but a small light having the mildness of
gas."
To a reporter for one of the leading
New York dailies who had shadowed
him to Ansonia, Edison described a vi-
sion of a central station for electric light-
ing that he would create for all of New
York City. A network of electric wire
would deliver current for a myriad of
small household lights, unlike the daz-
zling arc lights made by Farmer and
Brush. In some way electric current
would be metered and sold. Edison said
he hoped to have his electric-light in-
vention ready in six weeks! At Menlo
Park, N.J., where his already famous
workshop was located, he would wire all
the residences for light and hold a "grand
exhibition."
Thus from the beginning Edison riv-
eted his attention not so much upon the
search for an improved type of incan-
descent filament as upon the analysis of
the social and economic conditions for
which his invention was intended. As he
turned with immense energy to expand-
ing the facilities at Menlo Park and se-
curing the essential financing, he con-
tinued his studies of the gas-lighting in-
dustry. In parallel he projected the eco-
nomics of the electric-lighting system he
envisioned.
Cas had its inconvenience and dan-
gers. "So unpleasant . . . that in the new
Madison Square theater every gas jet is
ventilated by small tubes to cany away
the products of combustion." But what-
ever is to replace gas must have "a gen-
eral system of distribution— the onlv pos-
sible means of economical illumination."
Gathering all the back files of the gas
industry's journals and scores of volumes
bearing on gas illumination, he studied
the operations and habits of the indus-
try, its seasonal curves and the layout of
its distribution systems. In his mind he
mapped out a network of electric-light
lines for an entire city, making the
shrewd judgment: "Poorest district for
light, best for power— thus evening up
whole city." He meant that in slum dis-
tricts there would be higher demand for
small industrial motors. Against tables
for the cost of converting coal to gas he
calculated the cost of converting coal
and steam into electric energy. An ex-
pert gas engineer, whose services Edison
engaged at this time, observed that few
men knew more about the world's gas
business than did Edison.
Edison had a homo oeconomicus with-
in him, a well-developed social and com-
mercial sense, though he was careless of
money and was not an accountant of the
type exemplified by his contemporary
John D. Rockefeller. Before the experi-
mental work on his invention was under
way, he had formed a clear notion,
stated in economic terms, of what its ob-
ject must be This concept guided his
search and determined the pattern of his
technical decisions, so that the result
would be no scientific toy but a product
useful to people everywhere. By his ini-
tial calculation of the capital investment
in machinery and copper for a whole
system of light distribution he was led
to define the kind of light he sought and
the kind of generating and distributing
system he needed.
Backers of the Electric Light
In the crucial matter of financing his
inventive work Edison had the generous
and imaginative aid of Crosvenor Low-
rey, a patent and corporation lawyer
well established in the financial com-
munity of Wall Street. Lowrey had fall-
en completely under Edison's spell and
regarded him much as a collector of
paintings regards a great artist whose
works he believes are destined for im-
mortality. Using his extensive connec-
tions and the favorable press-notices that
he encouraged Edison to secure during
late September and early October, 1878,
Lowrey assembled a sponsoring syndi-
cate of some of the most important fin-
anciers of the time. The underwriters
of the Edison Electric Light Company,
which was incorporated in mid-October,
included William H. Vanderbilt and
J. P. Morgan's partner Egisto Fabbri.
This was an unprecedented develop-
ment in U. S. business. Inventors had
been backed in the development of in-
ventions already achieved; Edison's fi-
nanciers were backing him in research
that was to lead to a hoped-for invention.
In many respects the venture marks the
beginning in this country of close rela-
tions between finance and technology.
"Their money," Edison said, "was in-
164
The Invention of the Electric Light
EDISON AND HIS PHONOGRAPH were photographed in 1878
by Mathew Brady. He had worked with electric lights but had
turned to the more promising phonograph. In the year that this
photograph was made, however, he resumed his work on lighting.
165
MENLO PARK was depicted in Frank Leslie's Illustrated News-
paper for January 10, 1880. The barnlike "tabernacle" of Edison's
laboratory is visible a! ihe far right. In its windows passengers on
the nearby railroad could see his experimental lights burning.
vested in confidence of my ability to
bring it back again." The 31-year-old
Edison was by now a well-known figure
in Wall Street. His quadruplex telegraph
system, by which four separate messages
could be transmitted over a single wire,
had furnished the pivotal issue in the vast
economic war waged between Western
Union and the rival telegraph empire of
the robber baron Jay Could. Edison's
carbon microphone had transformed the
telephone from an instrument of limited
usefulness to an efficient system of long-
range communication that was now ra-
diating across the country. The shares of
gas-lighting enterprises had tumbled on
the New York and London exchanges
upon Edison's announcement, in the
press campaign instigated by Lowrey,
that he was now about to displace gas
with electricity in the lighting of homes
and factories.
The alliance between Edison and his
sponsors was nonetheless an uneasy one.
The first rift appeared before the end of
October, when the rival inventor Wil-
liam Sawyer and his partner Albon Man
announced that they had "beaten" Edi-
son and applied for a patent on a carbon-
pencil light in a nitrogen-filled glass
tube. There was a flutter of panic in the
directorate of the Edison Electric Light
Company. The suggestion was made
that Edison should join forces with Saw-
yer and Man. Lowrey passed the sug-
gestion on to S. L. Criffin, a former
junior executive at Western Union whom
Lowrey had hired to help Edison with
his business affairs.
Criffin sent back a hasty "confiden-
tial" reply: "I spoke to Mr. Edison re-
garding the Sawyer-Man electric light.
... I was astonished at the manner in
which Mr. Edison received the informa-
tion. He was visibly agitated and said it
was the old story, that is, lack of confi-
dence. . . . No combination, no consoli-
dation for him. I do not feel at liberty to
repeat all he said, but I do feel impelled
to suggest respectfully that as little be
said to him as possible with regard to
the matter."
In view of Edison's talent for candid
and salty language Griffin's reticence is
understandable. After that there was no
further talk of consolidation with Saw-
yer or any other inventor.
The Menlo Park Laboratory
In his belief that he would "get ahead
of the other fellows" Edison was sus-
tained by his unbounded confidence in
his laboratory, its superior equipment
and its staff. The Menlo Park laboratory
was still the only full-time industrial re-
search organization in the country, in
itself perhaps Edison's most important
invention. During this period the physi-
cal plant was greatly expanded; a sep-
arate office and library, a house for two
80-horsepower steam engines, and a
glass blower's shed were added to the
original barnlike "tabernacle." Even
more important, Edison had collected a
nucleus of talented engineers and skilled
craftsmen, who were of inestimable help
to him in working out his ideas.
The self-taught Edison thought pri-
marily in concrete, visual terms. When
he was at work on the quadruplex tele-
graph, he had even built a model made
up of pipes and valves corresponding to
the wires and relays of his system, and
with running water replacing the elec-
tric current, so that he could actually see
how it worked. But now he would have
to depend far more on theory and mathe-
matics.
One of the happiest effects of Gros-
venor Lowrey 's personal influence was
the hiring of Francis R. Upton, a young
electrical engineer who had worked for
a year in the Berlin laboratory of the
great physicist Hermann von Helmholtz.
Edison jocularly nicknamed Upton "Cul-
ture," and, according to an oft-told story,
put the "green" mathematician in his
place with one of his scientific practical
jokes. He brought out a pear-shaped
glass lamp-bulb and gave it to Upton,
asking him to calculate its content in
cubic centimeters. Upton drew the
shape of the bulb exactly on paper, and
derived from this an equation for the
bulb's volume. He was about to compute
the answer when Edison returned and
impatiently asked for the results. Upton
said he would need more time. "Why, '
said Edison, "I would simply take that
bulb, fill it with a liquid, and measure
its volume directly!"
When Upton joined the staff late in
October, Edison had already committed
himself to the incandescent light. This,
rather than the arc light, was the way to
imitate the mildness of gas. But the fila-
ment glowing in a vacuum had been
sought in vain by numerous inventors
for half a century. In choosing the in-
candescent light rather than the arc-
light he was "putting aside the technical
advance that had brought the arc light
to the commercial stage." No one, in-
cluding himself, had succeeded in mak-
ing an incandescent lamp that would
work for more than a few minutes.
Edison's first efforts in 1878 were not
notably more successful. Knowing that
carbon has the highest melting point of
all the elements, he first tried strips of
carbonized paper as "burners" and man-
aged to keep them incandescent for
"about eight minutes" before they
burned up in the partial vacuum of his
glass containers. Turning to the infusi-
ble metals, he tried spirals of platinum
wire; they gave a brilliant light but
melted in the heat. Edison accordingly
devised a feedback thermostat device
that switched off the current when the
166
heat approached the melting point. The
lamp now blinked instead of going out
entirely. Nonetheless, with his eye on
the problem of financing, Edison filed a
patent application on October 5 and in-
vited the press in for a demonstration.
As this discouraging work proceeded
in the weeks that followed, Edison
turned, with Upton's help, to calculating
the current that would be consumed by
a lighting system equipped with a cer-
tain number of such lamps. They as-
sumed that the lights would be con-
nected in parallel, so their imaginary
householder could turn one light in the
circuit on or off at will, as in a gas-light-
ing system. Thinking in round numbers,
they assumed that these lamps, when
perfected, might have a resistance of one
ohm and so would consume 10 amperes
of current at 10 volts. Allowing in addi-
tion for the energy losses in the distribu-
tion system, they found that it would
require a fabulous amount of copper to
light just a few city blocks. Such a sys-
tem of low-resistance lights was clearly
a commercial impossibility.
This was the gist of the objections
which had greeted Edison's first an-
nouncements that he would use an in-
candescent bulb in a parallel circuit.
Typical of the scorn heaped upon him
was the opinion expressed by a commit-
tee set up by the British Parliament to
investigate the crash of gas-lighting se-
curities. With the advice of British sci-
entists, the members of the committee
declared that though these plans seemed
"good enough for our transatlantic
friends," they were "unworthy of the at-
tention of practical or scientific men."
From Ohm's law, which governs the re-
lationship between voltage, amperage
and resistance in a circuit, the report
argued that if an electric light of 1,000
candlepower were divided into 10
smaller lights and connected in parallel,
each of the smaller lights would radiate
not one tenth but "one hundredth only of
the original light." In this judgment such
figures as Lord Kelvin and John Tyndall
concurred. Before the Royal Institution
in London the distinguished electrician
Sir William Preece declared: "Subdivi-
sion of the electric light is an absolute
tgnts fatuus."
Ohm's law does indeed show that the
amount of current (amperes) flowing in
a circuit is equal to the electromotive
force (volts) divided by the resistance
(ohms) in the circuit. Edison's contem-
poraries reasoned that an increase in the
number of lights in a circuit would in-
crease the resistance and therefore re-
duce the flow of current to each. It was
thought that the only way to provide
these lights with sufficient current was to
reduce the resistance in the distribution
system. In a parallel circuit this meant
increasing the thickness of the copper
conductors to an impractical degree.
Such were the limits on the operation
The Invention of the Electric Light
of arc lights, with their low resistance
and huge appetite for current. Upton's
calculations showed that this conclu-
sion also applied to Edison's first low-
resistance incandescent lamps.
Edison now confounded his collabora-
tor by proposing that he make the same
sort of estimates for an entirely different
kind of circuit. This time he would
assume lights of very high resistance,
supplied with current at high voltage
and low amperage. In November and
December Upton made calculations on
the basis of the same number of lights,
but lights with the high resistance of
100 ohms each. These lights were to
operate on the low current of only one
ampere. Their high resistance was to be
offset, in accord with Ohm's law, by the
high voltage of 100 volts in the circuit.
The result was astonishing: A high-
resistance system would require only
one hundredth of the weight of copper
conductor needed for a low-resistance
system. And copper was the most costly
element involved— the decisive eco-
nomic factor.
The High-Resistance System
Here was the crux of Edison's insight
at Ansonia. He had recognized there
that the subdivision of light called for
lamps of high resistance which would
consume but little current; to balance
the electrical equation it would be neces-
INTERIOR OF EDISON'S LABORATORY at Menlo Park was
also depicted in the January 10, 1880, issue of Frank Leslie's
Illustrated Newspaper. At the time of the work on the electrir
light the laboratory had expanded into several other buildings.
167
sary to supply the current at high volt-
age. This was the "necessary secret" that
was "so simple." Today every high-
school physics student learns that the
power lost in transmitting electric ener-
gy varies with the square of the current.
Thus a tenfold reduction in current
meant a decrease of a hundredfold in the
energy wasted (or a hundredfold de-
crease in the weight of the transmis-
sion line). It was a conception easily
reached by an elementary applica-
tion of Ohm's law, but it had not oc-
curred to any of Edison's contem-
poraries. Even Upton did not immedi-
ately grasp the full import of Edison's
idea. As he said later: "I cannot imagine
why I did not see the elementary facts
in 1878 and 1879 more clearly than I
did. I came to Mr. Edison a trained man,
with a year's experience in Helmholtz's
laboratory, ... a working knowledge of
calculus and a mathematical turn of
mind. Yet my eyes were blind in com-
parison with those of today; and ... I
want to say that I had company!"
With Upton's figures before him Edi-
son was convinced that a new and strate-
gic invention lay surely within his grasp.
It was clear what kind of distributing
system he wanted. And he knew what
form of incandescent burner would serve
his purpose. To offer the necessary re-
sistance to the passage of current it must
have a small cross section and so would
have a small radiating surface.
By January, 1879, Edison was testing
his first high-resistance lamp. It had a
spiral of very fine platinum wire set in
a globe that contained as high a vacuum
as could be achieved with an ordinary
air pump. The results were encourag-
ing; these lamps lasted "an hour or two."
He then attacked the dual problem of
getting a higher vacuum and improving
his incandescing element. After another
trial with carbon, he returned to metals:
platinum, iridium, boron, chromium,
molybdenum, osmium— virtually every
infusible metal. He thought of tungsten,
but could not work it with existing tools.
Discouraged by the problem, Edison
tried nitrogen in his globe and then re-
sumed his efforts to obtain a higher
vacuum. Hearing of the new and effi-
cient Sprengel vacuum pump, which
used mercury to trap and expel air, he
sent Upton to borrow one from the near-
by College of New Jersey (now Prince-
ton University). When Upton returned
with the pump late that night, Edison
kept him and the other men on the staff
up the rest of the night trying it out.
GENERATOR which Edison developed for the needs of electric
lighting appears at right in this engraving from Scientific Ameri-
can for October 18, 1879 (at that lime this magazine appeared
weekly I. The generator was railed the "long-w .listed Mary Ann."
168
The Invention of the Electric Light
At this stage Edison made a useful
finding: "I have discovered," he noted,
"that many metals which have gas with-
in their pores have a lower melting point
than when free of such gas." With the
aid of the Sprengel pump he devised a
method of expelling these occluded
gases, by heating the element while the
air was being exhausted from the bulb.
The platinum wire within the bulb
thereupon became extremely hard and
could endure far higher temperatures.
Edison later said that at this stage he
"had made the first real steps toward the
modern incandescent lamp."
Meanwhile the spirits of his financial
sponsors had begun to droop. Their bril-
liant inventor, far from having achieved
anything tangible, was hinting plainly
that he needed more monev. The first
Brush arc lights were ablaze over lower
Broadway, and more were being in-
stalled elsewhere with impressive effect.
Edison's backers began to have serious
doubts as to whether he had pursued the
right course. To shore up their morale
Lowrey arranged to have Edison give
them a private demonstration.
In April, as one of Edison's associates
recalled it, 'They came to Menlo Park
on a late afternoon train from New
York. It was already dark when they
were conducted into the machine shop
where we had several platinum lamps in-
stalled in series." The "boss" showed his
visitors pieces of platinum coil he was
using in the lamps, pointed out the
arrangement of the lights and described
the type of generator he hoped to build.
Then, the room having grown quite
dark, he told "Honest John" Kruesi to
"turn on the juice slowly."
"Today, I can still see those lamps
rising to a cherry-red . . . and hear Mr.
Edison saying 'A little more juice' and
the lamps began to glow. 'A little more,'
. . . and then one emits a light like a
star, after which there is an eruption
and a puff, and the machine shop is in
total darkness. . . . The operation was
repeated two or three times, with about
the same results."
The platinum coils still consumed a
lot of power for the light they gave, and
they were costly and short-lived. The
temporary Wallace-Farmer dynamos
heated up badly, and were not powerful
enough to enable Edison to connect his
lamps in parallel. Edison admitted that
the system was not yet "practical."
It was a gloomy gathering that broke
up on that raw April evening. All of
Lowrey's abounding faith would be nec-
essary to rally, the spirits and funds of
Edison's despondent backers. Some
VACUUM PUMP used to remove air from lamp bulbs (lop center) was of a new type about
which Edison had read in a scientific journal. The man is holding a vessel of mercury.
rumors of the disappointing demonstra-
tion leaked out; the price of Edison stock
fell sharply, while that of gas-lighting
securities rose. "After that demonstra-
tion," Edison's associate relates, "we
had a general house cleaning at the labo-
ratory, and the metallic lamps were
stored away."
Edison now rallied his staff to efforts
on a much broader area of the front
"under siege." He followed three main
lines of investigation. One group he
detailed to the task of developing the
dvnarno to supply the constant-voltage
current required by his high-resistance
system. He set another group to pulling
down a still higher vacuum in the glass
bulbs. The third team, under his watch-
ful eye, carried out the series of experi-
ments in which 1,600 different materials
were tested for their worth as incan-
descent elements.
The "Long-Waisted Mary Ann"
To subdivide the electric current for
numerous small lights in parallel Edison
needed a dynamo which would produce
169
a higher voltage than any dynamo in ex-
istence, and which would maintain that
voltage constant under varying demands
for current from the system. Existing
dynamos were designed around the fal-
lacious notion, held by most electrical
experts, that the internal resistance of
the dynamo must be equal to the ex-
ternal resistance of the circuit. Through
study of battery circuits they had proved
that a dynamo could attain a maximum
efficiency of only 50 per cent. In 1877 a
committee of scientists appointed by the
Franklin Institute in Philadelphia had
been impressed to discover that the most
successful European dynamo, designed
by Zenobe Theophile Cramme, con-
verted into electricity 38 to 41 per cent
of the mechanical energy supplied to
it. The efficiency of the Brush dynamo
was even lower: 31 per cent. These ma-
chines and their theoretically successful
contemporaries all produced current at
a relatively low voltage.
Edison had concluded, however, that
he must produce a dynamo of reduced
internal resistance capable of generating
current at a high voltage. Such a ma-
chine would not only meet the needs of
his lighting system but would also con-
vert mechanical energy to electrical en-
ergy with far greater efficiency. As his
associate Francis Jehl recalled, Edison
said that "he did not intend to build up
a system of distribution in which the ex-
ternal resistance would be equal to the
internal resistance. He said he was just
about going to do the opposite; he
wanted a large external resistance and
a low internal resistance. He said he
wanted to sell the energy outside the
station and not waste it in the dynamo
and the conductors, where it brought no
profits." Jehl, who carried out the tests
J
SERIES CIRCUIT (top) requires thai a number of electric lights (circles) be turned on
or off at the Mine time by a single switch (break in circuit). Parallel circuit (bottom),
which was adopted by Edison, makes it possible to turn lights on or off one at a time.
EDISON'S LIGHT.
IJie Great Inventor's Triumph in
Electric Illumination.
A SCRAP OF PAPER.
It Makes a Light, Without Gas or
Flame, Cheaper Than Oil.
TRANSFORMED IN THE FURNACE.
Complete Details of the Perfected
Carbon Lamp,
FIFTEEN MONTHS OF TOIL.
Siory of His Tireless Experiments with Lamps,
Barters acd Generators.
SUCCESS IN A COTTON THREAD.
The Wizard's Byplay, with Bodily Pain
and Gold "Tailings."
HISTORY OF ELECTRIC LIGHTING.
The oar approach of ine first puoltc exhibition of
Sduon's long looked for electric light, uooaued to
take place on New Vuri Eve at Menlo Pari, on
»UicU occasion that place will bo Illuminated with
the Dew light, has revived public lotereet In the
great inventor's work, and throughout tte mvlllsed
world scientists and people generally are anxiously
'•raiting the result. From the beginning of hie ex-
periments in electric lighting to the preeent lime
klr. Edison h.e kept his laboratory gaardedly
closed, and no authoritative account (except that
PubUahed lu the Uuild some months ago rotating,
to his first patent) of any of the important steps of
his progress has been made public — a course of pro-
cedure the Inreutor found absolutely necessary for
his own protection. The BxaiLO la now, however,
suabled to preeent to Its res dire a foil and accurate
account of his work from Its Inception to its oom-
pUilon.
a uoertZD nru.
EJitou's electric light, lucredible as It may appear.
Is produced from a little piece of paper — a tiny strip
or pauar that a breath would blow away. Through
FIRST NEWSPAPER ACCOUNT of Edi-
son's brilliant success appeared in The
/Vein York Herald for December 21, 1879.
170
The Invention of the Electric Light
of resistance, also remarked that the art
of constructing dynamos was then as
mysterious as air navigation. All elec-
t :cal testing was in the embryonic stage.
"There were no instruments for measur-
ing volts and amperes directly: it was
like a carpenter without his foot rule."
Upton himself had his difficulties in
this hirherto unexplored field: "I re-
member distinctly when Mr. Edison gave
me the problem of placing a motor in
circuit, in multiple arc, with a fixed re-
sistance; and I . . . could find no prior
solution. There was nothing I could find
bearing on the [effect of the] counter-
electro-motive force of the armature . . .
and the resistance of the armature on the
work given out by the armature. It was a
wonderful experience to have problems
given me by him based on enormous ex-
perience in practical work and applying
to new lines of progress."
The problem of a constant-voltage
dynamo was attacked with the usual
Edisonian elan. Seeking to visualize
every possible structural innovation for
his dynamo armature, he had his men
lay out numerous wooden dummies on
the floor and wind wire around them,
spurring them on in their task by laying
wagers as to who would finish first,.
After Edison had decided upon the
form of winding and type of electromag-
nets to be used, Upton made drawings
and tables from which the real armatures
were wound and attached to the com-
mutator. Edison eventually worked out
an armature made of thin sheets of iron
interleaved with insulating sheets of
mica; this armature developed fewer
eddy currents and so produced less heat
than the solid armature cores then used.
When the new cores were test-run, it
was Upton who made the mathematical
calculations from these tests and drew
up the final blueprints.
The self-effacing Upton can be given
principal credit for interpreting Edi-
son's ideas and translating them into
mathematical form. A careful student of
contemporary electrical knowledge, he
seems to have been conversant with, and
to have guided himself by, the design of
a German dynamo, made by the Siemens
works, that employed an auxiliary source
of current to excite its field magnets.
The new Menlo Park dynamo com-
prised many admirable features for that
period. With its great masses of iron and
large, heavy wires, it stood in bold con-
trast to its contemporary competitors.
Owing to the two upright columns of its
field electromagnets, it was nicknamed
"Edison's long-waisted Mary Ann."
When the dynamo was run at the cor-
rect speed, the voltage between its arma-
ture brushes was approximately 110,
and remained fairly constant, falling but
slightly when increasing amounts of cur-
rent were taken out of the machine.
Edison and Upton also contrived a sim-
ple but ingenious dynamometer by
which the torque of a drive belt was
used to measure the work output of the
steam engine that powered the dynamo.
When Kruesi completed the first oper-
ating machine, Upton carefully checked
the results. To his astonishment— and
quite as Edison had "guessed"— the new
dynamo, tested at full load, showed 90-
per-cent efficiency in converting steam
power into electrical energy.
Ldison was as jubilant as a small boy.
As was usual with him, the world was
soon told all about his "Faradic ma-
chine." It was described and depicted in
Scientific American for October 18,
1879, in an article written by Upton.
Once more there was scoffing at
Edison's "absurd claims." The hectoring
of Edison by some of the leading U. S.
electrical experts, among them Henry
Morton of the Stevens Institute of Tech-
nology, now seems traceable to their
ignorance. Reading Morton's predictions
of failure, Edison grimly promised that
once he had it all running "sure-fire,"
he would erect at Menlo Park a little
statue to his critic which would be eter-
nally illuminated by an Edison lamp.
As a matter of fact, this allegedly
ignorant "mechanic" was to be found
reading scientific journals and institu-
tional proceedings at all hours of the day
and night. It was thus that he had
learned about the Sprengel vacuum
pump. This device enabled him to
achieve an increasingly greater vacuum
and to test a broad variety of metals, rare
earths and carbon compounds under
hitherto unexplored conditions.
The globe itself was also much im-
proved, by the inventor's own design,
after he had brought to Menlo Park an
artistic German glass blower named
Ludwig Boehm. Edison one day drew a
sketch of a one-piece, all-glass globe
whose joint was completely sealed, and
late in April, 1879, Boehm, working skill-
fully with hand and mouth, fashioned
it in the small glass blower's shed in back
of the laboratory.
"There never has been a vacuum pro-
duced in this country that approached
anywhere near the vacuum which is
necessary for me," Edison wrote in his
notebook. After months of effort he could
say exultantly: "We succeeded in mak-
ing a pump by which we obtained a
vacuum of one-millionth part of an
atmosphere."
In the late summer of 1879 he realized
with growing excitement that a key posi-
tion had been won. He had a dynamo
supplying constant high voltage, and a
tight glass globe containing a high vacu-
um. In his mind's eye he saw what might
be done with an extremely fine, highly
resistant incandescing substance under
these conditions. His state of tension is
reflected in the laboratory notebooks by
such exclamations as: "S - - - ! Glass
busted by Boehm!" All that remained for
him was to discover a filament that
would endure.
The Carbon Filament
In late August or early September—
about a year after he first took up his
search— he turned back to experiment-
ing with carbon, this time for good. The
rods of carbon he had tried earlier had
been impossible to handle, as he now
understood, because carbon in its porous
state has a marked propensity for ab-
sorbing gases. But once he had a truly
high vacuum and a method for expelling
occluded gases he saw that he might
achieve better results with carbon than
with platinum.
In a shed in back of the laboratory
there was a line of kerosene lamps always
burning, and a laborer engaged in scrap-
ing the lampblack from the glass chim-
neys to make carbon cake. But lampblack
carbon by itself was not durable enough
to be made into fine lamp filaments. Edi-
son and Upton had an-ived at the conclu-
sion that, given a 100-volt multiple-arc
circuit, the resistance of the lamps should
be raised to about 200 ohms; this meant
that the filament could be no thicker
than a 64th of an inch.
Through the summer months Edison
and his staff worked at the tantalizing
task of making fine reeds of lampblack
carbon mixed with tar. His assistants
kept kneading away at this putty-like
substance for hours. It seemed impos-
sible to make threads out of it; as an
assistant complained one day, the stuff
crumbled.
"How long did you knead it?" Edison
asked.
"More than an hour."
"Well just keep on for a few hours
more and it will come out all right."
Before long they were able to make
filaments as thin as seven thousandths
of an inch. Edison then systematically
investigated the relations between the
electrical resistance, shape and heat
radiation of the filaments. On October
7, 1879, he entered in his notebook
a report on 24 hours of work. "A spiral
made of burnt lampblack was even bet-
ter than the Wallace (soft carbon) mix-
171
hire." This was indeed promising: the
threads lasted an hour or two before
they burned out. But it was not yet
good enough.
As he felt himself approaching the
goal Edison drove his co-workers harder
than ever. They held watches over cur-
rent tests around the clock, one man
getting a few hours' sleep while another
remained awake. One of the laboratory
assistants invented what was called a
"corpse-reviver," a sort of noise machine
that would be set going with horrible
effect to waken anyone who overslept.
Upton said that Edison "could never
understand the limitations of the
strength of other men because his own
mental and physical endurance seemed
to be without limit."
The laboratory notebooks for October,
1879, show Edison's mood of anticipation
pervading the whole staff. He pushed on
with hundreds of trials of fine filaments,
so attenuated that no one could conceive
how they could stand up under heat.
Finally he tried various methods of
treating cotton threads, hoping that their
fibrous texture might give strength to
the filament even after they had been
carbonized. Before heating them in the
furnace he packed them with powdered
carbon in an earthenware crucible
sealed with fire clay. After many failures
in the effort to clamp the delicate fila-
ment to platinum lead-in wires, Edison
learned to mold them together with
lampblack and then fuse the joint be-
tween them in the act of carbonization.
Then, as Edison later related, it was
necessary to take the filament to the
glass blower's shed in order to seal it
within a globe: "With the utmost pre-
caution Batchelor took up the precious
carbon, and I marched after him, as if
guarding a mighty treasure. To our con-
sternation, just as we reached the glass
blower's bench, the wretched carbon
broke. We turned back to the main
laboratory and set to work again. It was
late in the afternoon before we produced
another carbon, which was broken by
a jeweler's screwdriver falling against
it. But we turned back again and before
nightfall the carbon was completed and
inserted in the lamp. The bulb was ex-
hausted of air and sealed, the current
turned on, and the sight we had so long
desired to see met our eyes."
"Ordinary Thread"
The entries in the laboratory note-
books, although bare and impersonal,
nonetheless convey the drama and sense
of triumphant resolution pervading the
laboratory that night: "October 21—
No. 9 ordinary thread Coats Co. cord
No. 29, came up to one-half candle and
was put on 18 cells battery permanently
at 1:30 A.M No. 9 on from 1:30
A.M. till 3 P.M.-13X hours and was then
raised to 3 gas jets for one hour then
cracked glass and busted."
As the light went out the weary men
EARLY EXPERIMENTAL LAMP is de-
picted in one of Edison's notebooks. Thi>
lamp had a filament of platinum. It mehed.
FRANCIS R. UPTON made invaluable calculations for Edison's system. An electrical en-
gineer who had studied with Hermann von Helmholtz, he was named "Culture" by Edison.
172
The Invention of the Electric Light
waiting there jumped from their chairs
and shouted with joy. Edison, one of
them recalled, remained quiet and then
said: "If it can bum that number of hours
I know I can make it bum a hundred."
Yet all the workers at Menlo Park— Edi-
son, Upton, Kruesi, Boehm and the rest-
were completely astonished at their suc-
cess. They had become accustomed to
laboring without hope. "They never
dreamed," as one contemporary account
put it, "that their long months ... of
hard work could be ended thus abruptly,
and almost by accident. The suddenness
of it takes their breath away."
For once Edison tried to be discreet
and keep his momentous discoveries a
secret until he could improve upon his
lamp filament. At length, after experi-
menting with various cellulose fibers,
he found that paper, in the form of tough
Bristol cardboard, proved most endur-
ing when carbonized. Edison was ex-
ultant when this filament burned for 170
hours, and swore that he would perfect
his lamp so that it would withstand 400
to 1,000 hours of incandescence before
any news of it was published.
On November 1, 1879, he executed a
patent application for a carbon-filament
lamp. Its most significant passage was
the declaration: "The object of the in-
vention is to produce electric lamps giv-
ing light by incandescence, which lamps
shall have high resistance, so as to allow
the practical subdivision of the electric
light. . . . The invention consists in a
light-giving body of carbon wire ... to
offer great resistance to the passage of
the electric current, and at the same
time present but a slight surface from
which radiation can take place." The
specifications called for a distinctive
one-piece all-glass container, lead-in
wires of platinum that passed through
the glass base and were fused to the
carbon filament, and joints that were
sealed by fusing the glass.
Here were the essential features of
the basic Edison carbon-filament lamp,
in the form that was to be known to the
world during the next half century. It
was not the "first" electric light, nor
even the first incandescent electric
lamp. It was, however, the first practical
and economical electric light for uni-
versal domestic use.
Edison had spent more than $42,000
on his experiments-far more than he
had been advanced by his backers. Now
he asked for more money so that he
might complete a pilot light-and-power
station at Menlo Park: But the directors
were still uncertain about the future of
the invention. Was it "only a laboratory
toy," as one of them charged? Would
it not need a good deal of work before it ,
became marketable? Crosvenor Lowrey
stoutly defended his protege\ He got no
results until he prematurely, and over
Edison's objections, made the secret of
the electric lamp public.
Rumors had been spreading for sev-
eral weeks. New Jersey neighbors told
of brilliant lights blazing all night at
Menlo Park, and railroad passengers be-
tween New York and Philadelphia also
saw the bright lights with astonishment
from their train windows. In Wall Street
there was a flurry of speculation in
Edison stock; the price rose briefly to
$3,500 a share.
Then came a front-page story in The
New York Herald on Sunday, December
21, 1879. There followed an exclusive
article about the inventor's struggles for
the past 14 months, told to the world,
con amore, by Marshall Fox, who had
written much of Edison before. The de-
tailed treatment of such an adventure
in applied science as a feature story was
something of an innovation. Also some-
what unusual in the journalism of the
time was its relative accuracy of detail,
owing to help provided by Upton, who
also supplied drawings for the Herald's
Sunday supplement. The writer did his
best to explain how this light was pro-
duced from a "tiny strip of paper that
a breath would blow away"; why the
paper filament did not burn up but be-
came as hard as granite; and how the
light-without-flame could be ignited—
without a match— when an electric cur-
rent passed through it, giving a "bright,
beautiful light, like the mellow sunset
of an Italian autumn."
In the week following Christmas hun-
dreds of visitors made their way to the
New Jersey hamlet. Edison hurried with
his preparations for an announced New
Year's Eve display as best he could, but
was forced to use his whole staff of 60
persons to handle the crowds. He could
do no more than put on an improvised
exhibition, with only one dynamo and
a few dozen lights.
The closing nights of the year 1879
turned into a spontaneous festival that
reached its climax on New Year's Eve,
when a mob of 3,000 sight-seers flooded
the place. The visitors never seemed to
tire of turning those lights on and off.
The inventor promised the sight-seers
that this was but a token of what was
in store. He was awaiting the completion
of a new generator, he said, and intended
to illuminate the surroundings of Menlo
Park, for a square mile, with 800 lights.
After that he would light up the dark-
ness of the neighboring towns, and even
the cities of Newark and New York.
173
Hi-fi is a field in which erroneous ideas abound.
Both human and electronic factors are involved in
the accurate reproduction of sound.
11 High Fidelity
Edgar Villchur
Two chapters from his book Reproduction of Sound published in 1962.
It might appear that following a dis-
cussion of the nature of sound, the
logical subject to consider would be
the criteria for reproducing this sound
with "high fidelity" to the original. One
other element, however, should be cov-
ered first— the way in which we hear.
Perception of Sound
We have already seen, in examining
units of measurement for pitch and
power— the octave and the decibel— that
our perception of sound does not neces-
sarily correspond directly to the objec-
tive reality. The illusion is consistent,
however, so that a given sound always
has the same effect on a normal ear.
An important element in the percep-
tion of sound was discovered by Fletcher
and Munson in 1933. These investigators
demonstrated that our impression of
loudness did not depend solely on the
amplitude of the sound wave, but on
other things as well. Specifically, they
showed that sound in the lower treble
range of the frequency spectrum-the
3500-cps region— appeared to be much
louder than sound of the same amplitude
at any other part of the spectrum. Thus,
if the frequency scale was swept by a
tone which continuously rose in fre-
quency but kept exactly the same ampli-
tude, the loudness, or apparent ampli-
tude, would increase to a maximum at
about 3500 cps and then fall off again.
This fact does not have much practical
interest for the person listening to re-
produced music, except as it describes
the relative nuisance value of different
types of noise. No matter how lop-sided
our interpretation of acoustic reality, we
make the same interpretation in the con-
cert hall as in our living room, and the
craftsmen who designed musical instru-
ments (who worked to satisfy their ears,
not sound-level meters) perceived sound
in the same way.
Fletcher and Munson made a second
discovery, however, that does bear di-
rectly on the reproduction of sound.
They found that the effect described
above took place in varying degree, de-
pending on the over-all level of the
sound. For very high amplitude sound
the drop in loudness with frequency
below 3500 cps hardly occurred at all,
175
while for very soft sound the effect was
maximum. Above 3500 cps the effect re-
mained constant, within 2 or 3 db, no
matter what the over-all sound level.
The well-known "equal loudness con-
tours," also referred to as the Fletcher-
Munson curves, are reproduced in Fig.
2—1. Each curve plots the sound ampli-
tude required to produce the same per-
ceived loudness at different frequencies
of the scale. It can be seen that normal
hearing losses in the bass end become
progressively greater as the over-all
sound level is decreased.
This means that if an orchestra plays
a musical passage at the sound level rep-
resented by 90 db, and if this music is
reproduced at the 60 db level, we will
hear the bass with less relative loudness
than we would have heard it at the con-
cert itself. If you follow the 90- and
60-db curves, shown superimposed in
Fig. 2—2, you will see that there is ap-
proximately a 14 db perceived loss at
50 cps— it takes 14 db more of actual
amplitude, in the lower curve, to pro-
duce the same relative loudness at 50
cps as it does in the upper curve.
In order to re-create the original bal-
ance of perceived frequencies at low vol-
ume levels, it has become customary to
introduce bass boost which is related to
the setting of the volume control, either
automatically or otherwise.
A volume control tied to automatic
bass boost is called a loudness control.
(Some loudness controls also boost the
treble spectrum appreciably at low vol-
ume settings. There is no justification
for this in the Fletcher-Munson curves. )
High Fidelity to What?
The assumption will be made here
that the purpose of high fidelity equip-
ment is to reproduce as closely as possi-
ble the experience of the concert hall,
not to transcend or improve it.
I remember an exhibition at New
York's Museum of Modern Art, during
the late thirties, of "high fidelity" repro-
ductions of water color paintings. Life-
size reproductions were hung side by
side with the originals, and it was often
difficult or impossible to tell them apart.
There was no question in anyone's mind
about how to judge the quality of these
prints. The only criterion was accuracy.
The public that visited the exhibit was
used to looking at paintings, and was
able to make an immediate comparison
Fig. 2-1. The Fletcher-
Munton equal loud-
ness contours. For
•ach curvo, the
height at any point
repre»enf» the found
amplitude required to
produce the •am*
• ubjective loudness
as at 1000 cps. (After
Fletcher and Munson)
FEELING
yx
s.
-'z.
90
'r
J,
a>
1
^>
60
\j
J,
o
S
^
^^c
70
1%
J,
ai 60
s^s
^
s^\
eo
F-^.
>
K
y
>
^
50
>— —
^'
/
40
■^~-
>'
/
J,
z
J
z
20
J
'0
■J
0
0
1
-1 ll
1
1
1
500 WOO 2000
IN CrCLES PER SECOND
5000 10,000 20,000
176
High Fidelity
•X
* .20
> S «!0
£ z
i! »
z
o-'o
20 DB
II
»
1 H—
3D6 «o
i
#
-3
I '
i
i
>
'
90
-*3
:z
t
t
20 ' ' ' • •
100 iooo 10
FREQUENCY IN CYCLES PER SECOND
2001
Fig. 2-2. The 60 and
90 db Fletcher-Mun-
ton curvet superim-
posed. The shaded
area represents the
difference in normal
hearing loss from one
sound level to the
other.
between the copy and the original. No.
one thought of the prints as entities in
themselves, with qualities independent
of the qualities of the originals.
This point of view does not always
hold in the field of high fidelity musical
reproduction. Only a minority of today's
high fidelity public are concert-goers.
Many have never attended a live con-
cert; they know the sound of the orches-
tra or of individual musical instruments
only as it is reported by amplifiers and
loudspeakers. They may know what they
like in reproduced sound, but they have
no way of evaluating the realism of
reproduction.
This partly explains why so much vari-
ation is tolerated in audio equipment.
The same record may sound very dif-
ferent when played through different
brands of equipment, each brand equally
acceptable in the market place. The
evaluation of high fidelity components
is popularly thought of as an entirely
subjective matter, like comparing the
tone of one violin to that of another
rather than like holding a facsimile up to
its original.
For similar reasons high fidelity dem-
onstrations such as the annual Hi-Fi
shows can get away with a lot of sound
that is startling but essentially non-mu-
sical. Some of the "reproduced" sound
that greets the show visitor is necessarily
unfamiliar because it has no live coun-
terpart. A harmonica blown up in vol-
ume to the dimensions of a theatre organ
is a new and different instrument. A
crooner whispering into a microphone
an inch away invents a new sound; his
unamplified voice is never heard in pub-
lic. A combination of Bongo drum,
chimes and electric guitar creates a tutti
which one may like or dislike, but for
which there is no equivalent in one's
memory to serve as a live standard.
Such sound can only be accepted as
a self-sufficient entity, like an old calen-
dar chromo. Any resemblance to five
music or to painting is purely coinci-
dental, and the science and /or art of
reproduction is not really involved.
High fidelity has undoubtedly in-
creased rather than decreased the ranks
of music lovers, and there are probably
more people than ever who are unim-
pressed with gimmick sound. Many de-
signers and manufacturers in the field
work only for naturalness of reproduc-
tion. The designer of integrity avoids
like the plague those exaggerations that
sometimes attract the novice— over-em-
phasized bass for "depth," over-em-
phasized mid-range for "presence,"
over-emphasized treble for "brilliance."
These distortions are more properly
called, respectively, boominess, nasality
or "honkiness," and harshness.
Many demonstrations are not, fortu-
nately, of the gimmick type, and use
musical material played at musical lev-
els. There have also been concerts staged
with live musicians, in which direct com-
parisons of reproduced sound to the
sound of the live instruments could be
made, in the same way that direct com-
177
parisons of prints to original paintings
were made at the Museum of Modern
Art. The live vs. recorded public concert
is one method of giving direction to
equipment designers and perspective to
high fidelity consumers. Although trans-
ferring concert hall atmosphere to the
home has special problems of its own,
success in creating an identity of sound
in the concert hall itself solves the major
part of the problem. Even more vital to
maintaining balance and perspective in
the high fidelity world is live concert
attendance.
We are now prepared to discuss the
technical standards of quality that may
be applied to a sound reproducing sys-
tem. There will be no dividing lines pro-
posed, at which low fidelity becomes
medium, high, or super.
Frequency Response
The frequency response of a sound re-
producing system, or of one of its com-
ponents, describes its relative handling
of parts of the input signal which differ
in frequency. "Handling" may refer to
electrical amplification, as in an ampli-
fier, to conversion of mechanical to elec-
trical energy, as in a pickup, or to con-
version from electrical to acoustical
energy, as in a loudspeaker.
There are two aspects of frequency
response: the range of frequencies han-
dled, and the uniformity with which the
unit or system responds to different fre-
quencies. Knowledge of the first of these
is useless without knowledge of the sec-
ond. Let us therefore pass over the ques-
tion of range for the moment, and deter-
mine what uniformity will be required
for the range we finally decide on.
Uniformity of Response
Although the trained ear can usually
perceive a change of sound level of a
db or less in test signals, the average
observer is probably less sensitive to a
change of sound level in a particular
frequency range of a musical passage.
Reproduction which remains constant
over its frequency range within one or
two db would thus probably be ade-
quate for perfect apparent fidelity, other
things being equal.
This standard can be met in amplifiers
without much difficulty, even at high
power levels. The best pickups are also
able to conform, but loudspeakers are
laggard in this respect.
The results of non-uniform reproduc-
tion are several. Undue volume in a par-
ticular section of the sound spectrum
can produce stridency or boominess as
opposed to natural musical sound. More
particularly, the existence of sharp peaks
in the response curve, usually repre-
senting a resonant condition, mean that
hangover or ringing may be present—
the speaker cone or section of cone will
continue to vibrate after the signal has
stopped. This is perceived as a "rain-
barrel" effect, a muddying up of the
sound and impairment of the distinct-
ness of the different instrumental voices.
Such an effect is also indicated when the
listener is unable to distinguish clearly
the pitch of low-frequency tones.
Another important effect of peaked
frequency response is the exaggeration
of unwanted noise components such
as turntable rumble or record surface
scratch. This effect was not given its due
recognition in the earlier days of high
fidelity, when the existence of rumble
and surface noise was proudly displayed
as evidence of extended frequency range.
The amount of surface noise in a
good quality modern LP record and the
amount of rumble from a good record
player are such that there will not be
much significant noise produced in a
system with uniform frequency response,
even though the frequency range be ex-
tended to the limits of the present state
of the art. In a comparison test con-
ducted recently between two tweeters,
the one which was able to reproduce
almost an octave more of treble (into the
inaudible region) showed a dramatic
178
High Fidelity
decrease of surface noise, due to its ex-
treme evenness of response. There was
no selective reproduction of discrete fre-
quency regions, and the switch to the
superior speaker produced a fuller, more
natural treble simultaneously with the
reduction in surface noise.
A similar situation exists with regard
to turntable rumble. A peaked system
whose response falls off rapidly below
60 cps may exhibit more turntable
rumble than a smooth system whose
full response extends an octave lower.
Tell-tale evidence of the existence of
peaked reproduction in the bass may be
gathered from listening to the reproduc-
tion of speech. The male speaking voice
ordinarily contains no sound compo-
nents whose frequency is below 100 cps,
and the reproducing system should give
no hint (by a boomy, resonant quality
in the voice) that it is also capable of
speaking in the tones of the double bass.
Range of Response
It is generally agreed among acoustics
authorities that the range of 40 to 15,000
cps is sufficient for perfect or near-per-
fect apparent fidelity in the reproduction
of orchestral music. The phrase "near-
perfect" is meant to imply that when
such a range has been achieved the de-
signer should direct his attention to inac-
curacies of reproduction more gross than
are associated with the frequency limita-
tions indicated.
For the pipe organ enthusiast, how-
ever, there is significant intelligence
(significant, that is, from the point of
view of the emotional impact of the
music) down to 32 cps or lower. 32.7
cps is three octaves below middle C rela-
tive to A-440, and is the lowest note of
the average pipe organ, although many
larger organs reach down an octave
lower. These low organ tones are distin-
guished by the fact that they contain
a strong fundamental component. The
lowest tones of the piano, on the other
hand, contain no fundamental energy
that significantly affects the quality of
the sound. Even though the lowest key
on the piano strikes 27.5 cps, response
down to this frequency is not required
for the reproduction of piano music.
Probably no characteristic of audio
components is so freely booted about
by advertising copywriters as frequency
range. Any numerical range of frequen-
cies listed is totally meaningless unless
accompanied by a description of the
decibel tolerance above or below refer-
ence that is being used, and, for a loud-
speaker, by a description of off-axis re-
sponse as well. A 3-in. speaker made
for portable radios will "respond" when
stimulated by a 30-cps signal— perhaps
by having its cone tear loose and fly out
into the air— and almost any speaker,
even a woofer, will make some kind of
sound when stimulated by a high-pow-
ered 15,000-cps signal. A frequency re-
sponse rating must mean something
more than that a signal of given fre-
quency makes a speaker move audibly,
or that it makes an amplifier show an
electrical output of some sort at its ter-
minals. It must mean that within a stated
frequency range, and, for power devices,
within a stated range of power, the fun-
damental output of a given device is
uniform to a stated degree.
Treble Dispersion
The on-axis response of a loudspeaker
may be very deceiving, because the
higher frequencies tend to be directed
in a beam which continually narrows as
the frequency is raised. Good sound dis-
persion must therefore be a qualifying
factor for any treble response curve.
A speaker which has relatively uni-
form treble output both on-axis and off-
axis (over a reasonably large solid angle
—perhaps 45 degrees in any direction
from the axis) will reproduce music with
a "spaciousness" that does not exist
when there is more concentrated beam-
ing of the treble. Furthermore, severely
attenuated off-axis response in the treble
179
means that the total sound power radi-
ated at treble frequencies is considerably
less than that implied by the on-axis
response curve. It is this total radiated
power, rather than the on-axis pressure,
that determines whether a speaker will
sound dull, natural, or over-bright in a
normally reverberant room.
Transient Response
Transient response refers to the ac-
curacy of reproduction of the wave
envelope, and is concerned with the
reproduction of attack and decay char-
acteristics of the sound. We have seen
that uniform frequency response predicts
the absence of ringing; if the steady-
state frequency response curve does not
have peaks, the reproduced sound will
die away just as in the original.
Consider, for example, the tone repre-
sented in (A) of Fig. 2—3. Perfect re-
production would produce an identical
wave form, differing perhaps only in
amplitude, while poor transient response
would be indicated by the hangover that
is apparent in (B). The continuation of
the reproduced signal after the original
END OF REPRO-
DUCED TONE
SHOWING
HANGOVER
Fig. 2-3. Poor transient response.
has ended may be compared to a color
smear on a reproduced painting.
Attack time involves the reproduction
of frequencies higher than the funda-
mental. Although a percussive tone may
have a low fundamental pitch, the fre-
quency components associated with its
steep attack characteristic may be very
high. Natural reproduction of a drum
beat through a two-way speaker sys-
tem may thus be accomplished by the
"woofer" handling the fundamental tone
and its proper decay, while the "tweeter"
contributes the sound components that
make up the sharp attack.
Harmonic and Intermodulatlon
Distortion
Reproducing devices have a charac-
teristic way of performing with less than
perfect accuracy. In addition to the fre-
quencies at which they are asked to
vibrate mechanically (or alternate elec-
trically) they introduce new modes of
oscillation of their own— and these new
frequencies are harmonics, integral mul-
tiples of the original frequency. This
inaccuracy is called harmonic distor-
tion. It is measured as the ratio of the
amplitude of the spurious harmonics to
the true signal, in per cent.
We have seen that harmonics of fun-
damental frequencies are produced in
any case by musical instruments. Yet
small amounts of harmonic distortion
produce very unpleasant effects. The
sound becomes harsh, unmusical; the
bass is wooden and the treble painful.
The primary reason for this is that
with harmonic distortion comes an at-
tendant evil— intermodulation distortion.
Intermodulation distortion can be de-
scribed as the introduction of new sound
components, at sum and difference fre-
quencies, when tones of two or more fre-
quencies are passed through a non-linear
system— that is, a system which creates
harmonic distortion. These sum and dif-
ference frequencies are harmonically un-
related to the original musical tones.
They are musically discordant, and they
serve to create raucous, unmusical sound
in a degree proportional to their relative
strength. The formation of intermodula-
tion products is illustrated in Fig. 2—4.
The primary importance of low dis-
tortion has always been recognized by
audio authorities. It has also become in-
180
NON-DISTORTING
REPRODUCING
DEVICE
SOLE COMPONENTS
ARE 50 CPS AND
1000 CPS
WWIMAM
50 CPS
REPRODUCING
DEVICE WITH
DISTORTION
WMAft/vwi
COMPONENTS
INCLUDE 950 CPS
AND 1050 CPS
1000 CPS
Fig. 2-4. Intermodulation distortion as a result of harmonic distortion of the low-frequency wave
form. Note that the wave envelope of the high-frequency tone is "modulated."
High Fidelity
creasingly recognized by the high fidel-
ity public in recent years, after the first
flush of excitement over reproducing
regions of the frequency spectrum pre-
viously untouched. Amplifier manufac-
turers now feature distortion data over
frequency response data; unfortunately
it is very rare for loudspeaker specifi-
cations to make any quantitative refer-
ence to distortion at all. The reason lies
in the fact that while both harmonic dis-
tortion and intermodulation distortion
(the latter is usually greater by a factor
of 3 or 4 ) can be kept to extremely low
values in high quality amplifiers— a small
fraction of one per cent at rated power—
the corresponding values for loudspeak-
ers are much higher. In the octave below
60 cps it is a rare speaker indeed which
can hold harmonic distortion, at any
appreciable sound level, below the 5 per
cent mark over the entire octave, and
many speakers produce percentages of
distortion in this frequency region ten
times as great. But the listening results
are not as bad as might appear at first
glance: speaker response is normally
severely attenuated in this lower range,
which helps, and there is comparatively
little musical material of such low fre-
quency to be distorted.
When the reproducing system has a
minimum of low frequency distortion,
very low bass tones of high power, such
as might be produced by organ pedal
pipes, not only remain pure in timbre
themselves but do not create intermodu-
lation with the rest of the music; they
do not destroy the purity of the treble
by introducing false tones.
Power Capability
The power capability of a high-quality
reproducing system should be such as
to be able to establish an intensity level
of sound in the living room equal to the
level at a good seat in the original con-
cert hall. The electrical power required
181
of the amplifier for achieving this goal
depends upon the efficiency of the
speaker, and the sound power required
of the speaker depends on the size and
other acoustical characteristics of the
room. Concert-hall level can be estab-
lished in a living room with a tiny
fraction of the acoustical power of a
symphony orchestra, because the lower
power is concentrated in a much smaller
area.
"Concert-hall level" is sometimes mis-
interpreted to mean the sound level
which would be created if the orchestra
were somehow jammed into the living
room itself. The writer has yet to ex-
perience at a live concert, even during
fortissimo passages, an assault on his
ears that compares to hi-fi assaults he
has weathered. It is interesting to note
that certain hi-fi demonstrations pre-
clude intelligible conversation which is
not shouted, while whispered conver-
sations in a concert hall are liable to
prove extremely distracting and annoy-
ing to one's neighbors. It is the sound
intensity level at the ear, not the power
of the orchestra, that we are trying to
reproduce.
Noise Level
Any sound component not present in
the original program material, other
than distortion products, is referred to
as noise, even though it may be periodic
and not conform to our strictly scientific
definition. Hum, rumble, surface scratch,
tube hiss or other circuit noise and sim-
ilar disturbances tend to destroy the
auditory illusion, and must be kept to a
minimum.
A standard for satisfactorily low noise
has been established by the FCC for FM
broadcast stations. It is that the power
ratio of the maximum signal to the noise
must always be at least 60 db; this rep-
resents a ratio of one million to one.
Dynamic Range
The dynamic range, or range of ampli-
tude of the reproduced sound from soft-
est to loudest, is determined by the two
factors just discussed, noise level and
power capability.
Soft musical passages can be masked
by any of the types of noise referred to,
and therefore the lowest sound levels
that can be used must be much louder
than the noise level. The maximum sound
levels that can be used, of course, are
limited by the power capability of the
system.
A dynamic range of 60 db, or a mil-
lion to one power ratio between highest
and lowest sound levels, is generally
considered adequate for reproduction
of the largest symphony orchestra.
Stereo
All of the above considerations apply
equally to monaural and to stereophonic
reproducing systems. These objective
elements of equipment fidelity— low dis-
tortion, adequate frequency response,
dynamic range, etc.— are able, in stereo,
to contribute more to the subjective illu-
sion of musical reality than in a mon-
aural system.
A stereo record-reproduce system
has in effect two parallel and complete
monaural systems. The work of each
component along the way is done twice.
The sound is picked up by two separate
microphones; the output of each micro-
phone is recorded on a separate track of
the tape; the record groove, although
not doubled, is cut in such a way as to
independently contain the record of
each signal channel; the pickup con-
tains two separate generating elements
which independently sense and trans-
mit each signal channel; the two signal
outputs of the pickup are sent through
independent amplifiers and fed to two
independent loudspeakers. There are
variations on this ideal scheme, but the
above describes the basic concept of
stereo.
The purpose of this dual-channel re-
production is, in the simplest terms, to
182
High Fidelity
help recreate the acoustical atmosphere
of the concert hall. In the old-fashioned
stereopticon each visual channel gave a
slightly different perspective view of the
subject. Similarly, in stereo recording,
each microphone gets a slightly differ-
ent auditory perspective. It is important
to note that this auditory perspective is
of the orchestra or soloists in the hall in
which they are performing, not merely
of the musical performers in the ab-
stract. This is important because a good
part of the sound that reaches our ears
at a concert does not come directly from
the orchestra, but is reflected from the
walls and ceiling of the concert hall.
The channels of a stereo system are
identified as "right and "left." This does
not mean that one microphone picks up
the sound of the right section of the
orchestra only, and that the other micro-
phone picks up the sound from the left
section of the orchestra. It does mean
that one microphone has a right-
oriented perspective of the total sound
in the recording hall, and that the other
microphone has a left-oriented perspec-
tive of the total sound. When these two
recorded channels (which, like the two
photos on a stereopticon card, are very
similar to each other) are reproduced
through two separate loudspeakers they
create, although not perfectly, the illu-
sion of the acoustical environment and
sense of space of the concert hall. There
is an increased awareness of the phys-
ical position of different instruments,
but this is very much less important
than the general increase in realism and
the consequent increase of clarity, par-
ticularly from the point of view of the
distinctness of the different musical
voices.
There is an approach to stereo
recording, commonly referred to as
"ping-pong" stereo, which provides an
exaggerated separation between the
right and left channels. If only the left
side of the orchestra were playing dur-
ing a particular passage, there would be
practically no sound from the right re-
cording channel. The left-right orienta-
tion of the different instruments is the
primary goal in this case, rathe; than
reproduction of the original acoustical
environment. The degree to which one's
attention is directed to the physical
position of the instruments in "ping-
pong" stereo is often much greater than
that at the live concert itself.
The greatest benefit of good stereo
recording and reproduction is that it
frees us, to a greater extent than was
possible previously, from the acoustical
environment of the listening room, and
transports us to some extent to the
acoustical environment of the hall in
which the recording was made. The
normal living room does not provide the
proper acoustical atmosphere for a
musical concert, particularly of a large
orchestra. Musical instrument designers
worked in terms of the tonal qualities
that would be produced in the type of
concert hall with which they were
familiar.
183
THE SOUND REPRODUCING SYSTEM
The phonograph is a classic example
of an invention that cannot be cred-
ited wholly to one man. In 1877 Edison
directed his assistant, John Kruesi, to
construct the first complete record-
reproduce system, but sound recorders
were sold on a commercial b?sis as early
as 1860, and Thomas Young's "A Course
of Lectures on Natural Philosophy" de-
scribed and illustrated a crude but prac-
tical sound recorder in 1807.
Young's recorder consisted of a sharp
metal stylus held by spring tension
against a revolving cylinder, the cylinder
coated with wax and turned by a gov-
ernor-controlled gravity motor. When
a vibrating body such as a tuning fork
was held against the stylus, a wavy
line was cut into the wax. This line rep-
resented the wave form of the vibra-
tions, and it could be studied and ana-
lyzed at leisure. The recorder was a
mechanical draftsman, that could sense
very small motions and record pressure
changes that took place within a period
of a very small fraction of a second.
By 1856 Leon Scott de Martinville
had constructed the "phonautograph"
(self-writer of sound) illustrated in
Fig. 3—1. The sound wave form was
scratched by a hog-bristle stylus on the
surface of a cylinder coated with lamp-
black, but the big advance over Young's
machine was the fact that the phonauto-
graph could record directly from the
air. The force of the acoustical vibrations
Fig. 3-1. The phonautograph of lion Scott do
Martinville — a commercial sound recorder of
the eighteen sixties. (Courtesy Smithsonian
Institution)
184
High Fidelity
was concentrated by a horn onto a dia-
phragm, and the stylus was attached to
the diaphragm, so that the recording
needle did not have to actually touch
the vibrating source of sound. This de-
vice, which corresponds in function to
the modern oscilloscope, was a catalogue
item of the Paris firm of Koenig, and was
sold as a measuring instrument to acous-
tical laboratories.
The phonautograph which is at the
Smithsonian Institution at Washington
would undoubtedly reproduce music if
a proper record were placed on its re-
volving cylinder. The theoretical possi-
bility of playback was understood then,
too, but the lampblack records were use-
less for playback, as their grooves were
not rigid enough to direct the vibrations
of a playback needle. About half a year
before Edison got his brainstorm Charles
Cros conceived a method for bringing
the groove sinuosities back to life as
sound. The lampblack recording was to
be photo-engraved on a metal cylinder,
and running a needle through the hard
groove would then cause the needle to
vibrate from side to side, in the same
time pattern as the hog bristle stylus
that first inscribed the line.
For reasons which may be related to
nineteenth century differences in tradi-
tion between the scholar and the indus-
trial engineer, Cros didn't even construct
a working model, but merely filed a com-
plete, sealed description of his system
with the Academie des Sciences. On the
other hand, less than a month after
Edison first conceived of a reproducing
phonograph the country was reading
about a working unit in newspaper head-
lines. There was a great stir of excite-
ment over this amazing tonal imitator,
(see Fig. 3-2) with public demonstra-
tions, lectures before august scientific
bodies, and a visit to the White House.
Rg. 3-2. Edit on with his tin-foil phonograph. (Photograph by Brady — courtesy Smithsonian Insti-
tution)
185
The excitement soon died down, as
the Edison machine was an impractical
toy, with neither permanent records nor
usable fidelity. The recorded groove was
indented into a semi-hard material, tin
foil; it was only able to retain its shape
partially, and that for very few play-
ings. Subsequent technical improve-
ments, however, made the phonograph
a popular device by the turn of the cen-
tury. It is curious that our modem re-
cording system, in which the record is a
mechanical copy of the original master,
is more closely related to Cros' system
than to Edison's. Emil Berliner, the
father of the moulded or cast record,
began his research work by successfully
carrying out Cros' proposals.
The Mechanical or "Acoustic"
Phonograph
It would be useful to consider the de-
sign of the non-electric phonograph, as
illustrated in (A) of Fig. 3-3. A better
insight can thereby be gained into the
~5
L_ 1 O Q O
Fig. 3-3. (A) The mechanical phonograph. (B)
The electric phonograph.
function of the various components of
a modern electronic system.
The wave forms frozen into the record
groove control the vibrations of the play-
back stylus when the groove is dragged
past the stylus by a revolving turntable.
These stylus vibrations, although they
contain a fairly large amount of me-
chanical energy, engage practically no
air, like the revolutions of a bladeless
electric fan. The needle is therefore
attached to a diaphragm, which vibrates
in sympathy with the stylus and has a
much larger surface area in contact with
the air of the room.
But even the reproducing diaphragm
doesn't get a sufficient bite of the air
for practical purposes. Therefore the dia-
phragm is placed at the narrow throat
of an acoustical horn, and the actual
usable sound emerges into the room
from the much larger mouth of the horn.
The system works somewhat as though
the diaphragm area were really that of
the horn's mouth.
It can be seen that all of the energy
radiated by the horn is taken from the
mechanical vibrations of the needle, and
the forces between needle and record
groove are necessarily great. This has
obvious implications for record wear,
but perhaps more important, the de-
mands for power placed on the "sound
box" or "speaker" (old-fashioned terms
for the needle-diaphragm-head assem-
bly) place a severe limitation on musi-
cal fidelity. High distortion and peaked
and severely limited frequency response
are to be expected.
The Phonograph Amplifier
The solution to this problem lies in
changing the function of the phono-
graph pickup, from the primary genera-
tor of sound power to a device which
controls an outside source of power. If
the power from the outside source is
made to oscillate in imitation of the
needle vibrations, two benefits can result:
1. The final output sound derived
186
High Fidelity
from the record groove can be much
louder.
2. The power demands on the
pickup itself are no longer heavy. The
pickup can be designed for quality
rather than loudness; the problems
of achieving uniform, extended fre-
quency response and low distortion
are considerably lessened. So, inci-
dentally, is the required weight on
the pickup and the grinding away of
the record groove.
The control of an outside source of
power to conform to given oscillations
is called amplification. The first phono-
graph amplifier was pneumatic: the
needle was made to actuate an air valve,
which periodically throttled a flow of
compressed air. Most of the work of ra-
diating sound power was thus performed
by the air compressor, and the stylus
was relieved of part of its burden.
All modern sound reproducing sys-
tems use amplifiers, but unlike the first
pneumatic systems these amplifiers are
electronic. The phonograph pickup is no
longer a sound generator but an elec-
tric generator. It produces small alter-
nating voltages at its terminals, whose
wave forms conform to those of the
groove and of the recorded sound. The
pickup has to generate very little power,
because the output voltage can be ampli-
fied to almost any desired degree. The
amplified electrical power must finally,
of course, be converted back into sound
by a loudspeaker. The two types of re-
producing system, electrical and purely
mechanical, are shown in Fig. 3—3.
The Modern Sound Reproducing
System
The purpose of the historical approach
used above has been to furnish the
reader with an appreciation of the rea-
son for the modern audio system being
designed as it is. With the electronic
amplifier supplying the brute force, so
to speak, the mechanical components-
pickup and loudspeaker— can be built
in such a way as to suppress the natural
resonant tendencies inherent in mechani-
cal vibratory systems.
Before discussing each of the audio
components in detail, it would be useful
to make a brief survey of the entire re-
producing system. A complete monaural
system is illustrated in Fig. 3—4.
First of all the disc record must be
revolved by a motor and turntable. The
chief operational requirements of this
part of the system are that it revolve at
the correct speed, that the speed be con-
stant, and that extraneous vibrations
do not communicate themselves to the
pickup.
The first of these requirements is for
the purpose of keeping the reproduced
music at the same absolute pitch at
which it was recorded: too fast a turn-
table speed will make the pitch sharp,
and too low a speed will make it flat.
The second condition listed, constant
speed, is required in order to avoid pitch
variations, or "wow." The third require-
ment, lack of extraneous vibrations,
keeps low-frequency noise called "rum-
ble" out of the final sound.
The groove variations are sensed by
the needle, or stylus, which in high-
quality systems is jewel tipped; it is
usually diamond. The needle must have
an unmarred, smooth surfaced, hard tip,
normally of spherical shape.
The pickup is an electric generator
(usually either of the piezo-electric,
variable reluctance, or moving-coil type)
whose function is to translate the me-
chanical vibrations of the needle into
electrical oscillations of the same wave
form. It must do this with minimum
distortion of the wave form, and must
not allow resonances of its own to in-
fluence its output voltage significantly.
It is also an advantage for the pickup
to impose as little work as possible on
the needle. The greater the force re-
quired for the groove to displace the
needle from side to side, the greater the
vertical bearing force will have to be to
187
TAPE MECHANISM
PICKUP
STYLUS
P
TONE ARM
1ST
MOTOR
I
a a a — [
itHnnQQ
CONTROL UNIT
POWER AMPLIFIER
SPEAKER
SYSTEM
Fig. 3-4. Diagram of a complete monaural •ound reproducing system.
maintain proper and constant stylus-
groove contact, and the greater the
wear of both record and needle.
The tone arm holds the pickup in
place over the groove, and must pro-
vide sufficient freedom of motion so
that the pressure of the groove walls
alone can make the needle move across
the record, following the recorded spiral.
It must also be free enough to follow
warp and eccentricity of the disc easily.
The tone arm must hold the pickup ap-
proximately tangent to the groove being
played, must provide the proper vertical
force for the pickup, and must not allow
its own resonant behavior to influence
the system.
The electrical output of one type of
pickup, the piezo-electric, is usually fed
directly to the amplifier. It is of the
order of V2 volt or more, and is a fairly
accurate replica of the recorded sound.
This is so because the characteristic fre-
quency response of the pickup is more
or less the inverse image of the frequency
characteristics "built in" to the record.
(This last subject will be taken up in
detail later.)
The reluctance and moving-coil pick-
ups, however, produce a much smaller
amount of electrical energy. The output
voltage of these pickups (which are
classed together as magnetic types) may
be as low as a few thousandths of a
volt. Furthermore the characteristic fre-
quency response of the magnetic pickup
does not compensate for the way in
which the frequency characteristics of
the recorded sound has been doctored.
Therefore the pickup output must be
passed through a preamplifier before it
enters the amplifier proper.
The preamplifier is normally com-
bined with the main amplifier control
sections (volume and tone controls).
Its functions are to increase the output
voltage of the pickup, and to compensate
accurately for the frequency character-
istics of the record so that the sound is
not deficient in bass and heavy in the
treble. Since different record companies
have made records with different char-
acteristics the preamplifier may allow
the operator to choose between several
types of frequency compensation. The
need for such control, which is called
variable record equalization, has disap-
peared with modem records, which are
standardized on the RIAA recording
characteristic.
The control section of the amplifier
allows the operator to regulate the vol-
188
High Fidelity
ume, and, in most cases, to either ac-
centuate or attenuate ("boost" or "cut")
the bass and treble portions of the repro-
duced sound independently. The pri-
mary function of tone control is to com-
pensate for deficiencies in associated
equipment or program material, and to
compensate for acoustical conditions of
the room in which the music is heard.
When the control section and phono-
graph preamplifier are combined on one
chassis, the entire unit is commonly re-
ferred to as a preamplifier.
The power amplifier receives the elec-
trical signal as it is finally shaped, and
releases another signal, ideally identical
in all respects except power. The power
amplification may be tens of millions
of times, from a fraction of a micro-
watt (one millionth of a watt) to dozens
of watts.
Although the demands on the ampli-
fier are very great, and although it
appears to be the most complicated of
the system components, it is the least
imperfect of these components. The per-
centages of harmonic and intermodula-
tion distortion, the irregularities of fre-
quency response, and the extraneous
noise introduced by an amplifier built
according to the best current design
practice, and without regard for cost,
are such that they are not limiting fac-
tors in the fidelity of the reproduced
sound.
The final component of the sound
system is the loudspeaker system, which
consists of the speaker mechanism itself
and the speaker enclosure. The loud-
speaker converts the alternating elec-
trical output of the amplifier into me-
chanical vibrations of a cone or dia-
phragm. But the cone vibrating by itself
cannot, for reasons that will be discussed
further on, produce adequate bass en-
ergy. It must be mounted in an enclo-
sure or baffle of some sort, which gives
the vibrating surface the "bite" of air
that it needs to radiate low-frequency
sound.
The speaker and its enclosure, like the
amplifier, should introduce as little dis-
tortion and frequency irregularity into
the signal as possible. Typical speaker
deficiencies are irregular frequency re-
sponse, poor transient response (hang-
over), and harmonic and intermodula-
tion distortion.
Two other components are shown in
Fig. 3—4. The tuner is a device which
converts AM or FM radio signals to
audio signals that can be handled by the
audio amplifier; the tape transport
mechanism, with its associated pream-
MUITIPLEX FM TUNER
Fig. 3-5. A stereo reproducing system.
189
plifier, provides a signal of the same cated); the stereo tuner receives the
nature as that coming from the tuner "multiplex" FM stereo signal and sepa-
or phonograph pickup. rates it into two separate channels,
Fig. 3-5 shows the basic elements of which it feeds independently to each of
a stereo reproducing system. The stereo the control units. Each control unit and
tape mechanism has two heads which each power output is shown duplicated,
independently reproduce each channel The two control units and power ampli-
that is recorded in parallel on the tape; fiers may be separate, or they may be
the stereo pickup provides two separate combined on one chassis, or all four
output signals from the two channels units may be combined on one chassis,
recorded in the groove (the turntable but in any case they must provide inde-
and pickup arm do not have to be dupli- pendent amplification for each channel.
190
Since the Niagara power plant was built, commercial
electric power has been almost entirely alternating
current. Now new consideration is being given to
the advantages of direct current for long distance
power transmission.
12 The Future of Direct Current Power Transm
ission
N. L Allen
A popular article published in 1967.
The history of technology provides many examples of
unexpected turns of fortune, and electrical technology is
no exception. It frequently happens that a principle or
technique, originally the basis of a well-established
system, is superseded by a device making a significant
advance, only to reappear in a different guise as the 'last
word' in the state of the art. An obvious example is the
crystal of the early radio receiver. This was superseded
by the thermionic valve, but it has now developed into
the more sophisticated form of the transistor. Not many
years before the era of the crystal receiver, an appreciable
proportion of electrical energy was generated, trans-
mitted, and used in the form of direct current. At that
time, generation and consumption usually took place in
the same locality, distribution was simple, and the
quantities of energy transmitted were small by modern
standards. However, serious limitations appeared as it
became necessary to distribute electrical energy more
widely, and direct current as the distributing medium
gave way to alternating current.
In many countries, the economic advantages of being
able to concentrate power generation in large stations
have led to the adoption of a comprehensive network of
power lines that interconnect generating plant and the
areas where the power is used. As the length of a power
line increases, the current passed, for minimum power
loss, decreases: the economic operating voltage for trans-
mission of a given power therefore increases. The trans-
mission of larger quantities of energy at high voltages
and low currents is greatly facilitated by the ease with
which alternating current can be transformed to the
voltage most appropriate for the power lines. In the
receiving areas of the system, the voltage can equally
easily be transformed to lower values suitable for distri-
bution, and a system of far greater flexibility can be set
up than is the case with direct current. Fur .her, it is
difficult to switch and, particularly, to interrupt direct
current. The interruption of an alternating current by
circuit breakers is relatively easy because the current
passes through zero twice in every cycle.
This combination of circumstances made alternating
current the natural choice as power systems increased in
size. The main links operated initially at 132 kilovolts,
but the need for increased power during the post-war
years has led to the adoption of 275 kilovolts and, more
recently, 400 kilovolts as the operating voltages of the
principal links in Britain. The power is distributed locally
at lower voltages. During this period, the remaining
direct current distribution systems have been reduced or
eliminated.
Transmission over long distances
What, then, is the place of direct current? There is
certainly no good reason for turning away completely
from alternating current distribution. But there have
always been some situations in power distribution prac-
tice in which direct current has distinct advantages over
alternating current, and it is worth while considering
what these situations are.
One basic factor in power system design is the need to
find the simplest and most efficient means of transferring
power from one point to another. Figure 1 (a) shows the
basic three-phase alternating current system and figure
1 (b) a favoured direct current system, which has positive
and negative polarities on the two lines, and is linked by
convenors to alternating current for generation at one
end and distribution at the other. In both cases, the
maximum voltage to earth is E, but for alternating
current, it is the root-mean-square value Ejy/ '2 that
determines the power transmitted. This is lEIA cos 9/^2,
where lA is the current in each conductor, lagging behind
the voltage in phase by 9 degrees. In the direct current
system, the power transmitted by each line is E/D, where
ID is the current. For transmission of equal power by the
two systems, therefore, it can be shown that each
alternating current line has 4/(3 cos* 9) times the cross
sectional area of the corresponding direct current line, a
factor which is always greater than 133. Moreover, the
191
alternating current system requires three cables rather
than two, so that the amount of copper required is
2/cos2 <f times that in the direct current system, a factor
which is always greater than 2.
Direct current, then, reduces the cost of the cables.
This may appear trivial compared with the other capital
costs in electrical systems, but over great distances, as in
the United States and the Soviet Union, the saving in
cable, and in the means of supporting the cable, becomes
a very significant factor that can outweigh the cost of
providing the convenor stations at each end of the
system.
Great distances bring further problems in alternating
current transmission that do not occur with direct
current. These problems arise from the relationship be-
tween the wavelength of the oscillation and the dimen-
sions of the system. The quarter-wavelength of a 50
cycles per second wave in air is about 900 miles, and the
transmission of energy through a conductor can be re-
garded as due to an influx of energy along its length from
the electromagnetic field that surrounds it. Over short
distances, this field is very nearly the same at all points,
since electromagnetic energy is conveyed with the
FT
_4.._
(b)
Figure 1 Simplified distribution systems: (a) alternating
current, (b) direct current.
velocity of light. But at distances greater than 900 miles,
the fact that the velocity of light is finite results in
significant differences, at any instant, in the phase of the
current at the two ends.
This situation leads to difficulties where two parts of a
power circuit, joined by a long alternating current link,
are out of phase and where a loop is formed through
another part of the network of different length. Large
circulating currents will be set up unless some form of
compensation is applied. A direct current link obviates
these difficulties; as a corollary, it may be noted also
that if a direct current line is used to link two alternating
current systems, they need not be synchronized with
each other.
Transmission over short distances
For long-distance transmission, overhead lines, supported
by towers, are used. The virtues of direct current are
most clearly shown when the current is carried by under-
ground or underwater cables. Here, the central core of
the cable, which is at the transmission voltage, is sur-
rounded by an insulant, the exterior of which is at earth
potential. This constitutes a coaxial capacitor, and the
capacitance per mile of a cable rated at 200 kV is
typically about 03 microfarads. In an alternating current
circuit, this capacitance is charged and discharged,
through the inductance and resistance of the cable itself,
once every half-cycle. Additional generating capacity is
needed to supply this charging current. In the example
quoted, at 200 kV, the charging current requires about
5000 ikilovolt-amperes per mile of cable ; at 400 kV
the figure is about 15 000 kilovolt-amperes per mile. For
appreciable lengths of cable, the losses become such that
the charging currents must be supplied at intermediate
points. At 200 kV, these points are about 25 miles apart
for 50-cycle alternating current; at 400 kV, only 15
miles. Thus, alternating current transmission becomes
impracticable in cables over long distances. Further, the
cost of the generating capacity needed to supply the
charging current is significant. Taking a rough figure of
£50 per kilowatt of installed capacity at the generating
station, this extra cost is £250 000 per mile for a 200
kilovolt cable. By contrast, with direct current in the
steady state, there is no charging current. It may well be
worthwhile, therefore, to accept the cost of converting
to direct current to avoid having to provide this charg-
ing current. Direct current is also advantageous in that
there are no dielectric losses due to reversal of the elec-
tric stress in the insulant.
The balance between the two systems
To summarize, direct current has significant advantages
for the transmission of bulk power over great distances
by overhead lines, and over short or long distances by
cable. In addition to the technical advantages already
examined, direct current may be valuable in linking two
alternating current systems that need not then be
synchronized. Alternatively, a very large alternating
current system mav be divided by direct current links
192
The Future of Direct Current Power Transmission
into two or more smaller systems: this is a possible future
development as power systems continue to increase in
size. It is necessary, however, to examine some disad-
vantages of direct current, and some relevant non-
technical factors, to demonstrate the balance affecting
the final choice of system.
The most obvious drawback to the use of direct
current is the need for conversion at each end of the link
in order to integrate it with established alternating
current systems. The technical details are outlined later,
but it may be mentioned here that the cost of the con-
version equipment is about twice that of the alternating
current equipment required for the termination of a
power line of corresponding size and output . These
costs must be set against the savings inherent in the
direct current system. There is therefore, a limit to the
length of a line, below which the capital outlay on a
direct current system is higher than that of an alternating
current system. Estimates of the critical length for a long
overhead line naturally vary, depending mainly on the
power to be transmitted and the voltage to be employed,
but figures of more than 300 miles have frequently been
quoted. This approach is unlikely to be favoured,
therefore, in the British Isles, but such systems are being
developed in the United States and in the Soviet Union.
For underground or submarine cables, where dielectric
losses and charging currents are so important, the
'critical length' is reduced to about 30 miles, and it is in
short submarine links and in urban transmission lines
that direct current finds its second important application.
Indeed, where large amounts of power have to be intro-
duced into large cities, legal and social considerations
may predominate over technical and economic factors.
It is frequently extremely difficult to obtain permission
to erect overhead lines in urban areas, and the distur-
bance to local amenities caused by the towers for high-
tension cables may not be justifiable. Underground
cables become necessary, and it is preferable to use
direct current for distances greater than about 30
miles.
In choosing between the systems, the fact that there
can be no direct current transformer and that there is no
satisfactory circuit breaker ensures that alternating
current maintains its general superiority for distribution
purposes. The use of direct current is thus confined to
the bulk transmission of high power between discrete
parts of a system or between two separate systems.
193
The Reader for Unit 3 contained the first part of
Newman's biography of this outstanding mathematician
and physicist. This final part covers primarily his
work on electromagnetic theory.
James Clerk Maxwell, Part II
James R. Newman
A biographical essay published in 1955.
In February, 1858, Maxwell wrote a letter to his aunt, Miss
Cay, beginning, "This comes to tell you that I am going to have
a wife." "Don't be afraid," he added, "she is not mathematical,
but there are other things besides that, and she certainly won't
stop mathematics." His engagement to Katherine Mary Dewar,
daughter of the principal of Marischal College, was formally
announced the same month, and in June they were married.
195
Their union became very close: they enjoyed doing things
together — horseback riding, reading aloud to each other,
traveling — and he even found useful tasks for her in his
experimental work. The marriage was childless, but this very
fact increased the couple's dependency and devotion. Maxwell
regarded the marriage tie in an "almost mystical manner."
The published letters to his wife overflow with religiosity.*
The Aberdeen appointment terminated in 1860 when the
two colleges, King's and Marischal, were fused into a new
university and Maxwell's chair in physics at Marischal was
eliminated. He was not long at liberty. In the summer of the
same year he became professor of natural philosophy at
King's College, London, a post he retained until 1865. The
teaching schedule at King's was long and arduous; in the
evenings there were lectures to be given to "artisans" as part
of his regular duties. Living in London offered him the oppor-
tunity to see something of Faraday, with whom, up to this time,
Maxwell had had only correspondence, to make the acquaint-
ance of other scientific men and to renew old friendships. He
was no solitary. "Work is good, and reading is good, but
friends are better," he wrote to his friend Litchfield.
Yet despite academic duties and social distractions, the five
years in London were the most productive of his life. The
paper "On the Theory of Three Primary Colors," the two
articles in the Philosophical Magazine on "Physical Lines of
Force" and the culminating electrical memoir "A Dynamical
Theory of the Electromagnetic Field," the Bakerian lecture
"On the Viscosity or Internal Friction of Air and other Gases,"
and the celebrated paper "On the Dynamical Theory of Gases,"
all belong to this period. He also performed important experi-
mental work during these years. At his house in Kensington,
* He did not write in this vein to others and it is a little puzzling why he found
it necessary in corresponding with her to quote Scriptures, to express the fer-
vent hope that the Lord would protect her from evil, and that she would get
her eyes off "things seen and temporal and be refreshed with things eternal."
196
James Clerk Maxwell, Part II
in a large garret, he measured the viscosity of gases and ob-
tained practical confirmation of the theoretical work I have
described. (For example, he found that the viscosity of air at
12 millimeters of mercury measured the same as at normal
pressure of 760 millimeters, thus proving that viscosity is in-
dependent of density.) To maintain the necessary temperature,
a fire had to be kept up in the midst of very hot weather and
kettles kept boiling to produce steam, which would be allowed
to flow into the room. Mrs. Maxwell acted as stoker. Another
investigation dealt with the ratio of the electromagnetic to the
electrostatic unit of electricity and led to one of Maxwell's
greatest discoveries. But I must postpone explaining this work,
even though to do so means abandoning the strict chronology
of events in Maxwell's life, until I have sketched the develop-
ment of his ideas on electricity.
To gain an appreciation of Maxwell's stupendous contribu-
tion to this branch of science it is necessary first to describe
very briefly the position of electrical theory when he embarked
on his studies.
In the eighteenth century, Charles Augustin de Coulomb
established the fundamental facts of electrostatic attraction
and repulsion. He showed that an inverse-square law, resem-
bling that of gravitational forces, applied to electric charges:
attraction or repulsion between charged bodies is directly
proportional to the product of the charges and inversely pro-
portional to the square of the distance between them.* (The
same discoveries, and others going beyond them, were made
earlier by the brilliant English recluse Henry Cavendish, but
his researches remained unpublished until 1879.) The next
major advance was that of Hans Oersted, who in 1819 found
that the flow of electric current through a wire parallel to a
magnetic needle makes the needle swing to a position at right
* F = k-^jr-, where F equals the force; k, a constant; q and q', the charges;
r2
r, the separating distance.
197
angles to the current. In other words, a current produces a
magnetic field.
A complementary series of advances was made early in the
same century by the French physicist and mathematician
Andre Ampere, whom Maxwell called the Newton of electric-
ity. The accolade was not undeserved, but there is a special
reason for Maxwell's conferring it: Ampere was the first to
explain the relationship of electric currents in terms of me-
chanical action,* an approach later perfected by Maxwell
himself. By experiment Ampere learned that a coil of wire
carrying an electric current behaves like a magnet, and he
suggested that a magnet owes its property to tiny electrical
currents inside the steel molecules. Thus a great conceptual
link was forged; for magnetism was shown to be not distinct
from electricity, but rather a name we give to some of the
effects of moving electric currents.
The crown of these fundamental researches was the im-
mortal work of Michael Faraday. He found that an electric
current flowing in one circuit can cause ("induce") a current
to flow in another circuit; that there is a magnetic field between
two currents; that a current can also be induced to flow in a
wire by use of a magnet — in other words, as a symmetric
counterpart to the phenomena discovered by Oersted and
Ampere, that changes in a magnetic field produce an electric
current.
Faraday's explanation of these phenomena is of central
importance to understanding Maxwell's work. He imagined
lines of force running through space as the instrumentality of
electric and magnetic actions.
These lines, it should be emphasized, were not conceived as
mere mathematical makeshifts, but as entities possessing phys-
ical properties. The lines spread out in every direction from
an electric charge or magnetic pole; every electric line of force
* He showed how to calculate the mechanical forces between circuits carrying
currents, from an assumed law of force between each pair of elements of the
circuit.
198
James Clerk Maxwell, Part II
starts from a positive charge and ends on a negative charge;
the more powerful the source, the more lines emanate from it.
Along the lines there is tension, a kind of pull, so that each
line behaves like an elastic thread trying to shorten itself; lines
of force repel each other sideways; the ends of a line of force,
representing charges, can move freely over the surface of a
conductor but are anchored on an insulator.
This hypothetical system, for which Faraday was convinced
he had found experimental evidence, was the starting point of
Maxwell's studies. He believed in it; he sought to develop it.
However, it must not be supposed that everyone accepted
Faraday's hypothesis. In fact, the majority of electricians — I
use the term in its older sense — regarded lines of force as a
concept much inferior to that of "action at a distance." They
likened electricity to gravitation. They imagined a charge (or
mass) situated at one point in space mysteriously influencing
a charge (or mass) at another point, with no linkage or con-
nection of any kind, however tenuous, bridging the distance
between the charges (or masses). Where Faraday sought to
assimilate the behavior of electricity to that of a mechanical
system, in which all parts are joined by levers, pulleys, ropes
and so on, the others held electricity to be a special case, to
which mechanical analogies were inapplicable. Maxwell ad-
mirably summarized the cleavage between the two views:
"Faraday, in his mind's eye, saw lines of force traversing all
space, where the mathematicians saw centres of force attract-
ing at a distance; Faraday saw a medium where they saw noth-
ing but distance; Faraday sought the seat of the phenomena in
real actions going on in the medium, they were satisfied that
they had found it in a power of action at a distance impressed
on the electric fluids."
Maxwell's first electrical paper "On Faraday's Lines of
Force" was delivered at Cambridge in 1855, within a few
months after he had finished reading Faraday's Experimental
Researches. What he tried to do was imagine a physical model
embodying Faraday's lines, whose behavior, like that of any
199
machine, could be reduced to formulae and numbers. He did
not suggest that the model represented the actual state of things;
on the other hand, he had no confidence in what mathematical
manipulations alone would reveal about the actual state of
things. It was important, he said, so to balance the method of
investigation that the mind at every step is permitted "to lay
hold of a clear physical conception, without being committed
to any theory founded on the physical science from which that
conception is borrowed."* Such a method will neither lead
* The opening paragraph of the paper is worth giving in full. "The present
state of electrical science seems peculiarly unfavorable to speculation. The laws
of the distribution of electricity on the surface of conductors have been analyt-
ically deduced from experiment; some parts of the mathematical theory of
magnetism are established, while in other parts the experimental data are want-
ing; the theory of the conduction of galvanism and that of the mutual attrac-
tion of conductors have been reduced to mathematical formulae, but have not
fallen into relation with the other parts of the science. No electrical theory can
now be put forth, unless it shows the connection not .only between electricity at
rest and current electricity, but between the attractions and inductive effects of
electricity in both states. Such a theory must accurately satisfy those laws the
mathematical form of which is known, and must afford the means of calculat-
ing the effects in the limiting cases where the known formulae are inapplicable.
In order therefore to appreciate the requirements of the science, the student
must make himself familiar with a considerable body of most intricate mathe-
matics, the mere attention of which in the memory materially interferes with
further progress. The first process therefore in the effectual study of the science,
must be one of simplification and reduction of the results of previous investiga-
tion to a form in which the mind can grasp them. The results of this simplifica-
tion may take the form of a purely mathematical formula or of a physical hypoth-
esis. In the first case we entirely lose sight of the phenomena to be explained;
and though we may trace out the consequences of given laws, we can never ob-
tain more extended views of the connections of the subject. If, on the other
hand, we adopt a physical hypothesis, we see the phenomena only through a
medium, and are liable to that blindness to facts and rashness in assumption
which a partial explanation encourages. We must therefore discover some meth-
od of investigation which allows the mind at every step to lay hold of a clear
physical conception, without being committed to any theory founded on the
physical science from which that conception is borrowed, so that it is neither
drawn aside from the subject in pursuit of analytical subtleties, nor carried be-
yond the truth by a favorite hypothesis. In order to obtain physical ideas with-
out adopting a physical theory we must make ourselves familiar with the exist-
ence of physical analogies. By a physical analogy I mean that partial similarity
between the laws of one science and those of another which makes each of them
illustrate the other. Thus all the mathematical sciences are founded on rela-
tions between physical laws and laws of numbers, so that the aim of exact sci-
ence is to reduce the problems of nature to the determination of quantities by
operations with numbers."
200
James Clerk Maxwell Part
into a blind alley of abstractions, nor permit the investigator
to be "carried beyond the truth by a favorite hypothesis."
Analogies are, of course, the lifeblood of scientific specula-
tion. Maxwell gives a number of examples, among them the
illuminating suggestion of William Thomson comparing the
formulae of the motion of heat with those of attractions (such
as gravitation and electricity) varying inversely as the square
of the distance. To be sure, the quantities entering into heat
formulae — temperature, flow of heat, conductivity — are
distinct from a quantity such as force entering into the formu-
lae of inverse-square attraction. Yet the mathematical laws of
both classes of phenomena are identical in form. "We have
only to substitute source of heat for center of attraction, flow
of heat for accelerating effect of attraction at any point, and
temperature for potential, and the solution of a problem in
attractions is transformed into that of a problem of heat."*
Maxwell proposed a hydrodynamical analogy to bring be-
fore the mind in "convenient and manageable form those math-
ematical ideas which are necessary to the study of the phe-
nomenon of electricity. "t The analogy was combined with
Faraday's lines of force, so that they were converted from
lines into "tubes of flow" carrying an incompressible fluid
such as water. He was then able to show that the steady flow of
particles of this fluid would give rise to tensions and pressures
corresponding to electrical effects. The fluid moving through a
system of such tubes represented electricity in motion; the
form and diameter of the tubes gave information as to strength
and direction of fluid (electric) flow. The velocity of the fluid
was the equivalent of electrical force; differences of fluid pres-
sure were analogous to differences of electrical pressure or
potential. Since the tubes were flexible and elastic, and ar-
* "On Faraday's Lines of Force," Transactions of the Cambridge Philosophical
Society, vol. X, part I, included in The Scientific Papers of James Clerk Max-
well, op. cit.
t Ibid.
201
ranged so as to form surfaces — each tube being in contact
with its neighbors — pressure transmitted from tube to tube
furnished an analogy to electrical induction.
One of Faraday's key concepts deals with the effect on space
of lines of magnetic force. A wire introduced into ordinary
space remains inert; but if magnetic lines of force are intro-
duced into the space, a current flows through the wire. Faraday
explained this by saying that the introduction of the magnet
threw the space into an "electro-tonic state." This concept
could not be fitted into the hydrodynamical analogy; Maxwell
admitted that while he could handle Faraday's conjecture
mathematically, the electro-tonic state at any point of space be-
ing defined "as a quantity determinate in magnitude and direc-
tion," his representation involved no physical theory — "it is
only a kind of artificial notation."*
It was a wonderful paper, and Faraday, to whom Maxwell
sent a copy, appreciated how much it advanced the "interests
of philosophical truth." "I was at first almost frightened," he
wrote Maxwell, "when I saw such mathematical force made to
bear upon the subject, and then wondered to see that the sub-
ject stood it so well."t Other students, however, thought the
subject stood it not at all well. Electricity was mysterious
enough without adding tubes and surfaces and incompressible
fluids. But Maxwell, who had good training in being consid-
ered queer, went on with the task of extending Faraday's ideas.
The second great memoir, On Physical Lines of Force, a
series of three papers published in the Philosophical Magazine
in 1861 and 1862, was an attempt to describe a more elaborate
mechanism that would not only account for electrostatic effects
but also explain magnetic attraction and Faraday's concept of
* For a discussion of Maxwell's use of physical analogy, see Joseph Turner.
"Maxwell on the Method of Physical Analogy," The British Journal for the
Philosophy of Science, vol. VI, no. 23, November, 1955.
t Campbell and Garnett, op. cit., p. 519.
202
James Clerk Maxwell, Part II
electromagnetic induction. Again, Maxwell used a concrete,
mechanical image to exhibit and develop his theory.* For, as
he said, "scientific truth should be regarded as equally scien-
tific whether it appears in the robust form and vivid colouring
of a physical illustration or in the tenuity and paleness of a
symbolic expression."
In the new model a magnetic field is produced by the rota-
tion in space of what Maxwell called "molecular vortices."
These may be thought of as slender cylinders (Maxwell him-
self had a disconcerting way of modifying the image as he
went along) that rotate round the lines of magnetic force. The
lines, traced by the pattern of iron filings about a magnet, are
the axes of rotation of the cylinders; the velocity of rotation
depends on the intensity of the magnetic force. Two mechan-
ical effects are associated with the cylinders: tension in the
direction of the lines of force, and pressure, exerted in the
"equatorial" direction (i.e., lateral pressure), arising from
the centrifugal force produced by the rotating cylinders. Com-
bined, these effects mechanically reproduce magnetic phe-
nomena: magnetism is a force exerted both along the axis and
outward from the axis.
It may now be asked how this curious arrangement fitted in
with the known facts that an electric current produces a mag-
netic field, and changing magnetic forces produce an electric
current. Step by step Maxwell answers this question.
The first point to clarify is the structure of a uniform mag-
netic field. Maxwell supposed this to consist of a portion of
space filled with cylinders rotating at the same velocity and in
the same direction "about axes nearly parallel." But immedi-
ately a puzzle confronted him. Since the cylinders are in con-
tact, how can they possibly rotate in the same direction? For
* As Turner top. cit.) points out. Maxwell employed two analogies. One
bridged a stationary field and a solid under stress. The other is between elec-
tricity and fluid motion, "with its suggestion that Ampere's laws be modified to
satisfy the equation of continuity."
203
Model of an electromagnetic field used by Maxwell visualized "Molecular
vortices" rotating in space. In this illustration the vortices are slender
cylinders seen from the end. (Maxwell gave the cylinders a hexagonal cross
section to simplify the geometry of the model.) Between the vortices are
small "idle wheels." If a row of the idle wheels is moved from A toward B,
they cause the adjacent vortices to rotate in the opposite direction. ( Scientific
American)
as everyone knows, a revolving wheel or cylinder causes its
neighbor to revolve in the opposite direction; thus one would
expect the rotation of the cylinders to alternate in direction
from one to the next. Maxwell hit upon a pretty idea. He sup-
posed the cylinders to be separated by rows of small spheres,
like layers of ball bearings, which acted as gears (in Max-
well's words, "idle wheels"). This arrangement, resembling a
device envisaged a century earlier by John Bernoulli, the
204
James Clerk Maxwell, Part II
younger, fulfilled the requirement. The spheres rotate in an
opposite sense to that of each of the cylinders with which they
are in contact, and so the cylinders all rotate in the same direc-
tion.
And now, as just reward for his ingenuity, Maxwell found
that the spheres could be made to serve another, even more
valuable, purpose. Think of them as particles of electricity.
Then by purely mechanical reasoning it can be shown that
their motions in the machine of which they are a part serve to
explain many electrical phenomena.
Consider these examples. In an unchanging magnetic field
the cylinders all rotate at the same constant rate; thus they
maintain a constant magnetic field. The little rotating spheres
keep their position; there is no flow of particles, hence no elec-
tric current, a result that tallies with the properties of a uni-
form magnetic field. Now suppose a change in the magnetic
force. This means a change in the velocity of rotation of the
cylinders. As each cylinder is speeded up, it transmits the
change in velocity to its neighbors. But since a cylinder now
rotates at a slightly different speed from its neighbor, the
spheres between them are torn from their positions by a kind
of shearing action. In other words, they begin to move from
their centers of rotation, in addition to rotating. This motion of
translation is an electric current; again, a result that tallies
with the properties of a changing magnetic field.
Observe now how the model begins to live a life of its own.
It was designed, as J. J. Thomson has pointed out,* to exhibit
Faraday's great discovery that magnetic changes produce elec-
tric currents. It suggested to Maxwell the no less striking con-
verse phenomenon that changes in electric force might produce
magnetism. t Assume the spheres and cylinders are at rest. If
* Sir J. J. Thompson, "James Clerk Maxwell." in James Clerk Maxwell, A
Commemoration Volume, op. cit.
t Ampere, of course, had already demonstrated that currents in wires produced
accompanying magnetic fields.
205
a force is applied to the spheres of electricity, they begin to
rotate, causing the cylinders of magnetism with which they are
in contact to rotate in the opposite direction. The rotation of the
cylinders indicates a magnetic force. Moreover, the model
holds up even as to details. Take a single illustration. Mag-
netism acts at right angles to the direction of flow of current. If
you will examine the diagram of Maxwell's model, you will
see that the cylinders will rotate in the direction perpendicular
to the motion of the spheres, thus bearing out the observation
that a magnetic force acts at right angles to the flow of a cur-
rent.
"I do not bring it forward," Maxwell wrote of his system,
"as a mode of connection existing in Nature. ... It is, how-
ever, a mode of connection which is mechanically conceivable
and easily investigated, and it serves to bring out the actual
mechanical connection between the known electromagnetic
phenomena.* Certain aspects of these "mechanical connec-
tions" have already been discussed — rotations, pressures,
tensions — which account for the reciprocal relations between
currents and magnetic forces. t The connections also serve to
explain the repulsion between two parallel wires carrying cur-
rents in opposite directions, an effect produced by the centrifu-
gal pressures of the revolving cylinders acting on the electrical
particles between them. The induction of currents is similarly
elucidated: this phenomenon, says Maxwell, is simply "part of
the process of communicating the rotary velocity of the vor-
tices [cylinders] from one part of the field to another." In
other words, whenever electricity (Maxwell's particles) "yields
to an electromotive force," induced currents result. His dia-
gram and the accompanying text make this beautifully clear.
Maxwell was not done with his model. It had helped in the
* "On Physical Lines of Force," op. cit.
t The model explained, for example, why a current of electricity generated heat:
for as the particles (or spheres) move from one cylinder to another, "they
experience resistance, and generate irregular motions, which constitute heat."
206
James Clerk Maxwell. Part II
interpretation of magnetism, the behavior of electric currents,
the phenomenon of induction; it had yet to pass the supreme
test: that is, to supply a mechanical explanation of the origin
of electromagnetic waves. To orient ourselves in this matter we
must examine briefly the question of condensers and insulators.
An electric condenser is a device for storing electricity. In
its simplest form it consists of two conducting plates separated
by an insulating material, or dielectric as it is called. The
plates can be charged, after which the charges attract each
other through the dielectric and are thus said to be "bound" in
place. Faraday in his experiments had come upon a curious
fact. He found that two condensers of the same size, fed by the
same electric source and with insulation of equal thickness,
differed markedly in their capacity to take or to hold a charge
if the insulating material (dielectric) was different. This was
difficult to understand if all dielectrics were equally imperme-
able to an electric current. Moreover, if it were true, as Max-
well already was beginning to suspect, that light itself is an
electrical phenomenon, how could light pass through certain
dielectrics, empty space among them? With the help of his
model, Maxwell advanced a bold hypothesis. Conductors, he
said, pass a current when the electrical particles they contain
are acted upon by an electric force. Under such an impulsion,
the little particles move more or less freely from cylinder to
cylinder, and the current flows as long as the force persists.
Not so in a dielectric. The particles are present but an easy
passage from cylinder to cylinder is impossible. This fact may-
be taken as the characteristic attribute of a dielectric, having
to do with its physical structure. Yet it was known that "local-
ized electric phenomena do occur in dielectrics." Maxwell sug-
gested that these phenomena also are currents — but of a
special kind. When an electric force acts on a dielectric, the
particles of electricity are "displaced," but not entirely torn
loose; that is, they behave like a ship riding at anchor in a
storm. The medium in which they are located, the magnetic
207
cylinders, is elastic ; under pressure the particles move, a lim-
ited distance, until the force pushing them is balanced by the
stresses due to the elastic reaction of the medium. Thus a state
of equilibrium is attained. But as soon as the impelling force
ceases to act, the particles snap back to their original positions.
The net effect of these mechanical actions is twofold. First, the
initial displacement of the electric particles constitutes a cur-
rent that passes through the dielectric. A current of this type is
called a displacement current to distinguish it from currents
that flow through conductors and are therefore known as con-
duction currents.* Wherever there is an electric force, said
Maxwell, there is displacement; wherever there is displace-
ment, there is a current.
Second, whenever the pressure displacing the particles is re-
leased, and they snap back, they overshoot and oscillate briefly
about their fixed positions. The oscillation is transmitted
through the magnetic medium (the insulator) as a wave. This
wave is the return phase of the displacement current.t (Max-
well suggested this disturbance on analogy to the displacement
of an elastic solid under stress.)
Maxwell next arrived at an epoch-making conclusion. The
velocity of the displacement wave, or current, depends on the
electrical properties of the medium in which it moves. More-
over, this velocity, as he showed, was "within the limits of
experimental error, the same as that of light." Hence, he in-
* The contrast between displacement currents and currents through conductors
was vividly expressed by Henri Poincare. A displacement current, he said, is
an elastic reaction like the compression of a spring: it can only be effected by
pressure against resistance. Equilibrium is reached when resistance balances
pressure. When the pressure is removed the spring regains its original form. A
conduction current, on the other hand, is like a viscous reaction such as is en-
countered in moving a body immersed in water. It can be effected only by pres-
sure; the resistance depends on velocity; the motion continues as long as the
motive force acts, and equilibrium will never be established. "The body does not
return to the starting point, and the energy expended in moving it cannot be
restored, having been completely transformed into heat through the viscosity of
the water." (Maxwell's Theory and Wireless Telegraphy, New York, 1904.)
t If the electric force applied to the insulator is varied continually, it will pro-
duce a continually varying displacement wave: in other words, a continuing
current.
208
James Clerk Maxwell, Part
Electromagnetic wave as visualized by Maxwell is a moving disturbance
which tends to separate positive (plus sign) and negative (dot) charges. In
the drawing at the top, magnetic lines of force (arrows) lie at right angles
to the direction in which the disturbance is moving. The drawing at the bot-
tom depicts the two components of the electromagnetic wave. The electrical
component is shown in black, the magnetic component in color. (Scientific
American)
ferred, "the elasticity of the magnetic medium in air is the
same as that of the luminiferous medium, if these two coex-
istent, coextensive and equally elastic media are not rather
one medium."
209
More must be said as to how Maxwell actually arrived at
this conclusion. In the 1850s the German physicists Wilhelm
Weber and Friedrich Kohlrausch had investigated an impor-
tant relationship, namely, the ratios of electrostatic to electro-
dynamic action. The electrostatic unit of charge was defined
as the repulsion between two (like) unit charges at unit dis-
tance apart. The electrodynamic unit was defined as the repul-
sion between two definite lengths of wire carrying currents
"which may be specified by the amount of charge which travels
past any point in unit time." In order to compare the repulsion
between static charges with that between moving charges, a
factor of proportionality must be introduced, since the units
are different for static and dynamic phenomena. That is, one
must determine how many positive units of electricity flowing
in one wire, and negative units flowing in the other, are re-
quired to produce between the wires a repulsion quantitatively
equal to that between two static units. The factor turns out to
be a velocity; for since the length of the wires is fixed, and the
number of units of electricity passing a given point in a given
time can be measured, what the investigator must consider is
the dimension length divided by time or velocity. Weber and
Kohlrausch had found that the velocity of propagation of an
electric disturbance along a perfectly conducting wire is close
to 3 x 1010 centimeters per second. This was an astonishing
coincidence, for the figure was about the same as the velocity
of light as it had been determined a few years earlier by the
French physicist Hippolyte Fizeau.
Kirchhoff remarked the coincidence, but did not pursue it;
Maxwell did. In 1860 he attacked the problem experimentally,
using an ingenious torsion balance to compare the repulsion
between two static charges and two wires carrying currents.
The Weber-Kohlrausch results were roughly confirmed. Also,
at about the same time (he said, in fact, that the pencil and
paper work was done before seeing Weber's memoir), he cal-
culated the velocity of displacement currents in empty space
or in any other dielectric. The resulting values tallied closely.
210
James Clerk Maxwell, Part II
In other words, currents in a wire, displacement currents in a
dielectric, and light in empty space (which of course is a
dielectric) all traveled with the same velocity. With this evi-
dence at hand, which he communicated in a letter to Faraday
in 1861, Maxwell did not hesitate to assert the identity of the
two phenomena — electrical disturbances and light. "We can
scarcely avoid the inference," he said, "that light consists in
the transverse undulations of the same medium which is the
cause of electric and magnetic phenomena."
"On Physical Lines of Force," despite its cogwheels and
other gross mechanical adjuncts, may be regarded as the most
brilliant of Maxwell's electrical papers. If it did not claim to
give a picture of the true state of things, it was at least enor-
mously enlightening as to how electricity and magnetism could
interact in a purely mechanical relationship. Maxwell himself
summarized the achievements of the theory as follows. It ex-
plained magnetic forces as the effect of the centrifugal force
of the cylinders; induction as the effect of the forces called into
play when there is a change of angular velocity of the cylin-
ders; electromotive force as an effect produced by stress on the
connecting mechanism; electric displacement as a result of the
elastic yielding of the mechanism; electromagnetic waves as
an accompaniment of displacement. The paper is one of the
rare examples of scientific literature in which one may glimpse
the play of imagination, the actual exercise of inductive power,
the cultivation of nascent ideas.
None of the basic concepts unfolded in this memoir was
discarded in the more mathematical writings that followed.
But Maxwell now had to outgrow his model. In "A Dynamical
Theory of the Electromagnetic Field," published in 1864,*
Maxwell, in Sir Edmund Whittaker's words, displayed the
architecture of his system "stripped of the scaffolding by aid
of which it had first been erected."t The particles and cylinders
* Royal Society Transactions, vol. CLV.
t History of the Theories of Aether and Electricity: The Classical Theories,
London, 1951.
211
are gone; in their place is the field — "the space in the neigh-
borhood of the electric or magnetic bodies" — and the aether,
a special kind of "matter in motion by which the observed
electromagnetic phenomena are produced." The matter com-
posing the aether has marvelous properties. It is very fine and
capable of permeating bodies; it fills space, is elastic and is
the vehicle of "the undulations of light and heat." Yet for all
its refinements and subtleties, the medium is no less a mechan-
ical rig than the cylinders and spheres of its predecessor. It
can move, transmit motions, undergo elastic deformations,
store potential (mechanical) energy and release it when the
deforming pressures are removed. Though susceptible to the
action of electric currents and magnets, it is nonetheless a
mechanism that, as Maxwell said, "must be subject to the gen-
eral laws of Dynamics, and we ought to be able to work out all
the consequences of its motion, provided we know the form of
the relation between the motions of the parts." In the preceding
paper Maxwell already had devised a set of equations that
described the possible mechanical basis of electrical and mag-
netic phenomena, and, in particular, how certain changes in
electric and magnetic forces could produce electrical waves.
He now elaborated the hypothesis of displacement currents and
obtained the expressions that are in substance the famous Max-
wellian equations of the electromagnetic field.
In their most finished form the equations appear in the
Treatise on Electricity and Magnetism (1873), the culmina-
tion of Maxwell's researches, which he wrote at Glenlair in the
years following his resignation from King's College. This
celebrated work deals with every branch of electric and mag-
netic science and presents the results of twenty years of thought
and experiment. Maxwell remained faithful to Faraday, whose
point of view is emphasized throughout the Treatise. Charac-
terizing his own part as that of an "advocate," Maxwell makes
no attempt to describe the hypotheses propounded by Weber,
Gauss, Riemann, Carl and Franz Neumann, or Ludwig Lorenz,
212
James Clerk Maxwell, Part II
the foremost cultivators of the theory of action at a distance.
The task Maxwell set himself was, first, to formulate mathe-
matically electromagnetic phenomena as observed experi-
mentally, and, second, to show that these mathematical
relationships could be deduced from the fundamental science
of dynamics; or to put it another way, that the laws of elec-
tricity in motion could be derived from the laws applicable to
any system of moving bodies. As always, Maxwell was very
cautious in expressing himself about the nature of electricity.
It behaves, he said, like an incompressible fluid; "wherever
there is electric force there is electric displacement." These,
as J. J. Thomson observed, are the only definite statements
about electricity to be found in the treatise, which led Hertz
to say that Maxwell's theory is Maxwell's equations, and
caused Helmholtz to comment that "he would be puzzled to
explain what an electric charge was on Maxwell's theory be-
yond being the recipient of a symbol."
What are the Maxwellian equations? I cannot hope to give
an easy answer to this question, but at the cost of deliberate
oversimplification I must try summarily to explain them, for
they are the heart of the theory.
Maxwell based the equations on four principles: (1) that
an electric force acting on a conductor produces a current
proportional to the force; (2) that an electric force acting on
a dielectric produces displacement proportional to the force;
(3) that a current produces a magnetic force (i.e., a moving
electric charge is surrounded by a magnetic field) at right
angles to the current's lines of flow and proportional to its
intensity; (4) that a changing magnetic force (or field) pro-
duces a current proportional to the intensity of the force. The
third and fourth principles exhibit a striking symmetry. The
third is Faraday's law of electromagnetic induction, according
to which "the rate of alteration in the number of lines of mag-
netic induction passing through a circuit is equal to the work
done in taking unit electric charge round the circuit." Max-
213
MJL1_____.
Hgffl---
sliw
214
James Clerk Maxwell, Part
Lines of force appear in Electricity and Magnetism, left: "Uniform magnetic
field disturbed by an electric current in a straight conductor." above: "Two
circular currents." (Scientific American)
well's complementary law, the fourth principle, is that "the
rate of alteration in the number of lines of electric force pass-
ing through a circuit is equal to the work done in taking a unit
magnetic pole round it."
On this foundation two sets of symmetrical equations can be
erected. One set expresses the continuous nature of electric
and magnetic fields; the second set tells how changes in one
field produce changes in the other. In these formulations the
mechanical aspects of the theory are retained, perfect conti-
nuity is preserved by treating electricity as if it were an in-
compressible fluid, and wave phenomena are deduced as the
consequences of displacement in a dielectric.
How does the concept of the field enter the theory? We have
215
followed Maxwell as he stripped his model of its particles and
cylinders and reduced it to an aetherial medium. In the
Treatise, while not abandoning the medium altogether, he robs
it of almost all its attributes other than form. The matter of the
medium, as Poincare says, is left only with purely geometric
properties, the atoms dwindle to mathematical points, subject
to the laws of dynamics alone. The grin is left but the cat is
gone. It is a perfect example of mathematical abstraction.*
The aether is a thing that wiggles when it is prodded,
but does nothing on its own. An electromagnetic field con-
sists of two kinds of energy, electrostatic or potential en-
ergy, and electrodynamic or kinetic energy. The aether, like
* Einstein made an interesting comment about Maxwell's equations and his use
of the concept of the field. "He showed that the whole of what was then known
about light and electromagnetic phenomena was expressed in his well-known
double system of differential equations, in which the electric and the magnetic
fields appear as the dependent variables. Maxwell did, indeed, try to explain,
or justify, these equations by intellectual constructions. But he made use of
several such constructions at the same time and took none of them really seri-
ously, so that the equations alone appeared as the essential thing and the
strength of the fields as the ultimate entities, not to be reduced to anything
else. By the turn of the century the conception of the electromagnetic field as
an ultimate entity had been generally accepted and serious thinkers had aban-
doned the belief in the justification, or the possibility, of a mechanical explana-
tion of Clerk Maxwell's equations. Before long they were, on the contrary,
actually trying to explain material points and their inertia on field theory lines
with the help of Maxwell's theory, an attempt which did not, however, meet
with complete success. Neglecting the important individual results which Clerk
Maxwell's life work produced in important departments of physics, and con-
centrating on the changes wrought by him in our conception of the nature of
physical reality, we may say this: — before Clerk Maxwell people conceived of
physical reality — insofar as it is supposed to represent events in nature — as
material points, whose changes consist exclusively of motions, which are sub-
ject to partial differential equations. After Maxwell they conceived physical
reality as represented by continuous fields, not mechanically explicable, which
are subject to partial differential equations. This change in the conception of
reality is the most profound and fruitful one that has come to physics since
Newton; but it has at the same time to be admitted that the program has by
no means been completely carried out yet."
I am puzzled as to what Einstein meant in saying that Maxwell's equation
eliminated the notion of mechanism in explaining electromagnetic phenomena.
Similar views have been expressed by many other physicists and philosophers.
Maxwell himself would not have agreed with this position. His writings refute
it. The inference was drawn by his successors. But there is a more important
216
James Clerk Maxwell, Part II
a universal condenser, may be conceived as storing energy —
in which case, being elastic, it is deformed. Since the aether
fills all space and therefore penetrates conductors as well as
dielectrics, it no longer makes any difference whether we deal
with a conduction current or a displacement current; in either
case the aether is set in motion. This motion is communicated
mechanically from one part of the medium to the next and is
apprehended by us as heat, or light, or mechanical force (as
in the repulsion between wires) or other phenomena of mag-
netism and electricity. The ruling principle of all such phe-
nomena, it should be observed, is that of least action. This is
the grand overriding law of the parsimony of nature: every
point that requires clarification; namely, do the equations justify the inference?
It is true that a field is not the same as a material particle, and that the motion
of a particle is not the same as a change in a field. It is true also that the con-
cept "material particle" was long held to he intuitively clear, while the concept
"field" has never heen so regarded. This makes it easier to say mysterious
things about fields, which no one would dream of saying about particles. But a
more careful definition of these concepts, as physicists actually use them, raises
serious question as to whether a field is any less suited to a "mechanistic" ex-
planation than a system of material particles; indeed, whether a mechanistic
explanation fits either or neither case. In modern physics material particles are
not what they once were. They are pale abstractions, quite incapable of any-
thing so robust as a collision. But then what is a collision? One thinks of bil-
liard balls knocking together, as a pristine example. This, however, is a plain
man's way of thinking. The modern physicist has rid his mind of such seductive
images. (As far back as the eighteenth century, the Italian physicist Boscovich
proposed the idea that the heart of an atom is not solid substance but a mere
center of immaterial force.) As particles fade, the field becomes more substan-
tial. Properties are now ascribed to it that make it seem more real and more
potent than a billiard ball or a boulder. Of course the field is hard to describe
in homely terms. Yet it is quite capable, as physicists tell us, of doing homely
things. It produces and undergoes changes — now as if it were a cloud, now an
engine, now an ocean. In short it has mechanical effects. By this I mean effects
of a kind produced by what used to be called material particles. Moreover, it
has mechanical properties. By this I mean properties of a kind produced by
what we call a machine. The field can do things no system of particles or
machine yet conceived can do. Since it can also do all they can do, it is a super-
machine. Is there any point in saving the name? I think there is, to keep our
thinking straight. We ought to keep it to describe both fields and particles or
we ought to discard it entirely. If the word "mechanism" has any meaning in the
universe of refined observation, it has as much meaning in relation to fields as
to particles. At the same time I am quite prepared to believe that it has as little
meaning in one case as the other; for that matter, no meaning in either.
217
action within a system is executed with the least possible ex-
penditure of energy. It was of the first importance to Maxwell
that electrical phenomena should satisfy the principle, for
otherwise his mechanical explanation of the phenomena would
not have been possible.
With these points in mind, we may examine a set of Max-
well's equations in a form that describes the behavior of an
electromagnetic field under the most general conditions, i.e., a
field moving in empty space. No conductors are present, no
free charges, and the medium is a vacuum. The equations then
read
1) divE = 0
2) divH = 0
. _ 1 dH
3) curlE-- - -^
1 dE
4) curlH= - ^
The meaning of the symbols is as follows: E and H represent
electric and magnetic field strength; since they vary in time,
and from place to place, they are functions of the space co-
ordinates x, y, z (not shown) and of the time coordinate, t.
C is the velocity of light and enters the equations as the rate
of propagation; div (an abbreviation for divergence) and curl
(an abbreviation for rotation) represent mathematical opera-
tions whose physical meaning is explained below.
Divergence is essentially a measure of rate of change. In
words, then, equation 1
div E = 0
says that in a moving field the electric intensity is the same at
every point, i.e., the rate of change is zero at every point. More
loosely, this equation extends to the field the classical principle
that electric lines of force can be neither created nor de-
stroyed. Thus the equation says that the number of electric
lines of force, representing the field strength, that enter any
218
James Clerk Maxwell, Part II
tiny volume of space must equal the number leaving it. Mak-
ing use of still another analogy, if one conceives of electricity
in Maxwell's idiom, as an incompressible fluid, equation 1
states that as much fluid flows out of a tiny volume of space in
a given time as flows in.*
* For the reader interested in a little more detail, the following explanation
may be helpful. Equation 1 states that the divergence of the electric field inten-
sity is zero at any point in space and at any instant of time. The meaning of the
equation may be visualized as follows. It is customary to represent £ at a given
instant of time by a series of lines whose relative density in space is propor-
tional to E. These lines have direction because £ is a vector. Consider a point
P and a sphere surrounding P. Let us suppose that the intensity of the electric
field on the left hemispherical surface of the sphere is uniform over the surface
and is directed at each point perpendicular to the surface.
*
Q?
Suppose further that some change takes place in the electric field intensity E
in the region occupied by the sphere but such that on the right hemispherical
surface the field E is again uniform and perpendicular to the surface but strong-
er than on the left portion. We would indicate this increase in the intensity of
E by having more lines leave the sphere on the right than enter on the left.
Using the number of lines as a measure of E, we count the lines entering the
spherical surface and multiply this number by the area of the hemisphere, and
regard this product as negative. Let us next form the analogous product of the
area and the number of lines leaving the surface, and regard this product as
positive. The algebraic sum of these two products, that is, the positive plus the
negative, is called the net electric flux through the spherical surface. This net
flux is the divergence of E over the volume of the sphere. In our illustration
the net flux of E has increased as E passes through the sphere. Hence we
should say in this case that the divergence of E through the sphere is positive.
If we now divide this net flux through the sphere by the volume of the sphere,
we obtain the next net flux per unit volume. We now imagine that the sphere
becomes smaller and smaller and contracts to the point P. Of course the net
flux per unit volume changes and approaches some limiting value. This limiting
value, which is a mathematical abstraction, is div E at the point P. Thus div E
is essentially a measure of the spatial rate of change of E at the point P. Since
equation 1 says that for electric fields div E = 0 at each point P, we may say
that the net spatial rate of change of E is zero in empty space. More loosely
stated, this equation says that electric field lines are neither created nor de-
stroyed at the point P. It is to be noted that the phrase "spatial rate of change"
is intended to emphasize that the divergence is concerned with the way in
which E changes from point to point in space at the same instant of time. This
spatial rate must be distinguished from the rate at which some quantity, for
example, E itself in equation 4, may change during some interval of time.
219
Equation 2
div H = 0
makes the same assertion for magnetic lines as equation 1
makes for electric lines.
Equation 3
curl E — -55—
dt
is Maxwell's way of stating Faraday's law of induction. The
equation describes what happens in a changing magnetic field.
The right side expresses rate of change,—^ — , multiplied by a
1
very small factor, (the negative sign before the fraction
c
is purely a matter of algebraic convenience) ; the left side ex-
presses the fact that an electric field is created by a changing
magnetic field. But the equation is more than analytic; thanks
to the sign curl, it actually gives a picture of the event. A simple
diagram may help make this clear. Suppose the existence of a
magnetic field uniform over a region of space. We draw a
circle
_ H
surrounding a bundle of parallel lines, which represent the
intensity and direction of the magnetic field. The circle lies in
a plane perpendicular to the lines. If the field is changed (by
220
James Clerk Maxwell. Part II
motion or by increase or reduction of strength), it produces
an electric field that acts in a circle around the lines of mag-
netic force (though it may also act in other directions). By
summing the work done in moving unit electric charge around
the circle, we obtain what is called the net electromotive force
around the circle.* If the circle were made of wire, the chang-
ing magnetic lines would of course induce the flow of a cur-
rent; but even without a wire — and therefore no current — a
force would be induced. Dividing this force by the area en-
closed by the circle gives the net electromotive force (per unit
area) which "curls" around the circle. Now imagine the circle
growing smaller and smaller and shrinking finally to the point
P. By this limiting process we obtain a limiting value of the
net electromotive force per unit area : this is curl E at P. Thus
equation 3 says that the limiting value of electromotive force
per unit area equals the rate of change of H at the point P,
multiplied by the tiny negative fraction, J Or, again,
more loosely stated, a changing magnetic field creates an
electric field whose electromotive force per unit area at any
given point and instant of time equals the time rate of change
of the magnetic field at that point and instant.
Equation 4
curl H — — -^r
c Ot
says that, except for the change in algebraic sign (which has
to do with the directions of the fields), the roles of E and H in
* In physical terms, we obtain the net capacity of the electric field to move
current along the circle.
t The symbol c, which here stands for the ratio of the electrostatic to the elec-
tromagnetic units of electricity, is required to translate E (an electrostatic phe-
nomenon) and H (an electromagnetic phenomenon) into the same system of
units. The equation explains how Maxwell was able to connect electrical and
magnetic phenomena with the velocity of light, for c is in fact that velocity.
221
equation 3 may be reversed. At any given point and instant the
magnetomotive force (the analogue for magnetic fields of
electromotive force) per unit of area created by a changing
electric field is equal to the time rate of change of the electric
field multiplied by the tiny positive fraction — . Now, the
reader who has followed this discussion will perceive that the
time rate of change of E, -«— , is none other than Maxwell's
displacement current. For since the changes are taking place
in the dielectric known as empty space, the only currents that
can flow are displacement currents.* Prior to Maxwell, it was
thought that the magnetic field H could be produced only by
currents that flowed in wires passing through the circle. If no
wires were present, the law thought to be applicable was
curl H = 0. It was Maxwell's great discovery, deduced me-
chanically from his model and expressed mathematically in
this equation, that a time-varying electric field produces (or
must be accompanied by) a net "curled" magnetic force even
in an insulator or empty space. t
According to Maxwell's theory, the introduction of a time-
varying electric force in a dielectric produces displacement
waves with the velocity of light. To put it another way, it is the
surge and ebbing of the force that produces the periodic dis-
placement waves; a static charge would merely create an in-
stantaneous displacement, which would be fixed, but not a
* Equation 4 assumes the existence of this current and relates it quantitatively
to the magnetomotive force generated by the existent magnetic field. Physically
we may regard the magnetic field as creating the displacement current or, con-
versely, regard the displacement current as creating the accompanying magnetic
field and magnetomotive force.
t Maxwell called -=- the displacement current, the term "displacement" mean-
ing that the electric field intensity E was being altered or displaced as time
■qe
varies, and the term "current" suggesting that — had the properties of a cur-
ot
a ■ &E
rent Mowing in a wire even though — existed in empty space.
222
James Clerk Maxwell, Part II
wave. Now, an electric current, as we have seen, whether in a
dielectric or in a conductor, is accompanied by a magnetic
force; and similarly a periodic wave of electric displacement
is accompanied by a periodic magnetic force. The wave front
itself, as Maxwell showed, comprises electric vibrations at
right angles to the direction of propagation and a magnetic
force at right angles to the electric displacement. The com-
pound disturbance is therefore called an electromagnetic wave.
A light wave (which is a displacement wave) is, as Henri
Poincare later elaborated, "a series of alternating currents,
flowing in a dielectric, in the air, or in interplanetary space,
changing their direction 1,000,000,000,000,000 times a sec-
ond. The enormous inductive effect of these rapid alternations
produces other currents in the neighboring portions of the
dielectric, and thus the light waves are propagated from place
to place."
The electromagnetic theory of light was testable experi-
mentally, and indeed stood up remarkably well in laboratory
trials. But this was only a limited confirmation of Maxwell's
system, for if his reasoning was correct, there must be other
electrical waves produced by initial disturbances of differing
intensity. These waves would differ from light in wave length
and would therefore not be visible, yet it should be possible to
detect them with appropriate instruments. How to find them,
not to say generate them, was now the crucial problem. Max-
well did not live to see it solved. Not until ten years after his
death were his prophecies fulfilled and the skepticism of his
most distinguished contemporaries refuted. As late as 1888
Lord Kelvin referred to Maxwell's waves as a "curious and
ingenious, but not wholly tenable hypothesis" ; but a year later
Helmholtz's greatest pupil, Heinrich Hertz, nosed out Oliver
Lodge in the race to demonstrate their existence. In a series of
brilliant experiments he showed how electric waves could be
"excited" (i.e., generated) by oscillation and detected by
a circular conductor provided with a small gap; and how they
could be polarized, reflected, refracted, made to form shadows
223
and to interfere with each other. The connection, he said, "be-
tween light and electricity ... of which there were hints and
suspicions and even predictions in the theory, is now estab-
lished. . . . Optics is no longer restricted to minute aether
waves, a small fraction of a millimetre in length; its domain is
extended to waves that are measured in decimetres, metres and
kilometres. And in spite of this extension, it appears merely
... as a small appendage of the great domain of electricity.
We see that this latter has become a mighty kingdom."
The Treatise, written while Maxwell was "in retirement" at
Glenlair, drew only part of his energy. As a "by-work" during
the same period he wrote a textbook on heat, which appeared
in 1870, and a number of papers of considerable importance
on mathematics, color vision and topics of physics. He main-
tained a heavy scientific and social correspondence, enlarged
his house, studied theology, composed stanzas of execrable
verse, rode his horse, went on long walks with his dogs, visited
his r ghbors and played with their children, and made fre-
quent trips to Cambridge to serve as moderator and examiner
in the mathematical tripos.
In 1871 a chair in experimental physics was founded at
Cambridge. It is hard to realize that at the time no courses in
heat, electricity and magnetism were being taught there, and
no laboratory was available for the pursuit of these arcane
matters. The University, as a contemporary scholar delicately
observed, "had lost touch with the great scientific movements
going on outside her walls." A committee of the faculty began
to bestir itself, a report was issued, and the lamentable facts
fell under the gaze of the Duke of Devonshire, Chancellor of
the University. He offered the money for the building and
furnishing of the famous Cavendish Laboratory. Thomson, it
was known, would not leave his post at Glasgow to take the
new chair, and Maxwell, though at first reluctant to leave
Glenlair, yielded to the urging of his friends to offer himself
as a candidate. He was promptly elected.
224
James Clerk Maxwell, Part II
He now devoted himself to the task of designing and super-
intending the erection of the laboratory. His aim was to make
it the best institution of its kind, with the latest apparatus and
the most effective arrangements for research. He inspected
Thomson's laboratory at Glasgow and Clifton's at Oxford to
learn the desirable features of both and embody them in the
Cavendish. He presented to the laboratory all the apparatus in
his own possession and supplemented the Duke's gift by gen-
erous money contributions. With so many details to be taken
care of, the structure and its appointments were not completed
until 1874. The delay, while inevitable, was inconvenient. "I
have no place," wrote Maxwell, "to erect my chair, but move
about like the cuckoo, depositing my notions in the Chemical
Lecture Room in the first term, in the Botannical in Lent and
in the Comparative Anatomy in Easter." His "notions" were
the courses he gave, beginning in 1871, on heat, electricity and
electromagnetism, a schedule maintained throughout the ten-
ure of his chair. And though the audiences were often small,
some of the best students were soon attracted to his lectures,
which contained much important original work. The renais-
sance that followed in physical science at Cambridge was the
direct result of his influence.
Maxwell's classic Matter and Motion, "a small book on a
great subject," was published in 1876. About this time he
contributed articles on various subjects — "Atom," "Aether,"
"Attraction," "Faraday," among others — to the famous ninth
edition of the Encyclopaedia Britannica. His public lectures
include a charming discourse "On the Telephone," which,
though delivered when he was already very ill, is not only as
clear as his best expositions but filled with gay, amusing
asides. Speaking, for example, of "Professor Bell's inven-
tion," he comments on "the perfect symmetry of the whole
apparatus — the wire in the middle, the two telephones at the
ends of the wire, and the two gossips at the ends of the tele-
phones "A task that occupied him for five years, almost to
the very end of his life, was editing twenty packets of unpub-
225
lished scientific papers of Henry Cavendish, who was great-
uncle to the Duke of Devonshire. This splendid two-volume
work, published in 1879, did much to fix the reputation of an
immensely gifted investigator, whose important work on elec-
tricity was unknown to his contemporaries because the results
were confided only to his manuscripts. Maxwell repeated
Cavendish's experiments and showed that he had anticipated
major discoveries in electricity, including electrostatic capac-
ity, specific inductive capacity and Ohm's law.
As Maxwell grew older, friends remarked on his "ever-in-
creasing soberness" of spirit. This must not be taken to mean he
was invariably melancholy or withdrawn or that his nice sense
of fun — about himself no less than about others — had van-
ished. He continued to see his many friends, to write light
verse and parodies, to promenade with his dog Toby, who was
at Maxwell's side even in the laboratory, to play small prac-
tical, but never mean, jokes, to engage in what was called
"humorous mystification" by advancing preposterous scien-
tific ideas in conversation while keeping a straight face. All
things, he once remarked, are "full of jokes," though they are
also "quite full of solemn matters," and he was as likely to
stress their light as their grave aspect.
But it is true he became somewhat more reticent with the
passing years, and more and more concealed his feelings and
reflections beneath an ironical shell. The tough, rational,
Scotch common-sense cord of his nature had always been inter-
twined with threads of mysticism. Often plain, even blunt, in
his address, he also had an allusive way of speaking and
showed a fondness for parables. He had faith in science, yet
he was at bottom skeptical as to how much could be learned
from science alone about nature and meaning. It was all very
well, he felt, to have "ideal aspirations"; on the other hand,
"It's no use thinking of the chap ye might have been." His
contemporaries remember him as both modest and intellectu-
ally scornful, tentative in his scientific opinions and dogmatic
when others seemed to him to be immoderately self-assured.
226
James Clerk Maxwell. Part II
"No one knows what is meant by" so-and-so was his way of
answering a cocksure formulation of a scientific "truth."
The most striking of Maxwell's traits was his gentleness.
"His tenderness for all living things was deep and instinctive;
from earliest childhood he could not hurt a fly." An extraordi-
nary selflessness characterized his relationship to those close
to him. When his brother-in-law came to London to undergo
an operation, Maxwell gave up the ground floor of his house
to patient and nurse and left himself with a room so small that
he frequently breakfasted on his knees because there was no
room for a chair at the table. Mrs. Maxwell had a serious and
prolonged illness in the last years of Maxwell's life, and he
insisted on nursing her. On one occasion it is reported that he
did not sleep in a bed for three weeks. But his work went on
as usual, and he was as cheerful as if he enjoyed the ordeal —
which may indeed have been the case. Nor did he give the
slightest sign of being downcast or show self-pity when his own
fatal illness seized him.
In the spring of 1877 he began to be troubled with pain and
a choking sensation on swallowing. For some strange reason he
consulted no one about his symptoms for almost two years,
though his condition grew steadily worse. His friends at Cam-
bridge observed that he was failing, that the spring had gone
out of his step. When he went home to Glenlair for the sum-
mer of 1879, he was so obviously weakening that he called for
medical help. He was in terrible pain, "hardly able to lie still
for a minute together, sleepless, and with no appetite for the
food which he so required." He understood thoroughly that his
case was hopeless, yet his main concern seemed to be about the
health of his wife. In October he was told he had only a month
to live. On November 5 he died. "No man," wrote his physi-
cian, Dr. Paget, "ever met death more consciously or more
calmly." When Maxwell was buried in Parton Churchyard at
Glenlair, the world had not yet caught up with his ideas. Even
today it has not fully explored the kingdom created by his
imagination.
227
Oersted established a connection between electric
currents and magnetism; Faraday found the connection
between magnetic fields and induced electric cur-
rents. But it was Maxwell who synthesized and ex-
tended these two results.
14 On the Induction of Electric Currents
James Clerk Maxwell
An excerpt from his Treatise on Electricity and Magnetism
published in 1873.
528.] The discovery by Orsted of the magnetic action of an
electric current led by a direct process of reasoning to that of
magnetization by electric currents, and of the mechanical action
between electric currents. It was not, however, till 1831 that
Faraday, who had been for some time endeavouring to produce
electric currents by magnetic or electric action, discovered the con-
ditions of magneto-electric induction. The method which Faraday
employed in his researches consisted in a constant appeal to ex-
periment as a means of testing the truth of his ideas, and a constant
cultivation of ideas under the direct influence of experiment. In
his published researches we find these ideas expressed in language
which is all the better fitted for a nascent science, because it is
somewhat alien from the style of physicists who have been accus-
tomed to established mathematical forms of thought.
The experimental investigation by which Ampere established the
laws of the mechanical action between electric currents is one of
the most brilliant achievements in science.
The whole, theory and experiment, seems as if it had leaped,
full grown and full armed, from the brain of the 'Newton of elec-
tricity.' It is perfect in form, and unassailable in accuracy, and
it is summed up in a formula from which all the phenomena may
be deduced, and which must always remain the cardinal formula of
electro-dynamics.
The method of Ampere, however, though cast into an inductive
form, does not allow us to trace the formation of the ideas which
o-uided it. We can scarcely believe that Ampere really discovered
the law of action by means of the experiments which he describes.
We are led to suspect, what, indeed, he tells us himself*, that he
* Thdorie da rherwminet Electrodynamiqiut, p. 9.
229
discovered the law by some process which he has not shewn us,
and that when he had afterwards built up a perfect demon-
stration he removed all traces of the scaffolding by which he had
raised it.
Faraday, on the other hand, shews us his unsuccessful as well
as his successful experiments, and his crude ideas as well as his
developed ones, and the reader, however inferior to him in inductive
power, feels sympathy even more than admiration, and is tempted
to believe that, if he had the opportunity, he too would be a dis-
coverer. Every student therefore should read Ampere's research
as a splendid example of scientific style in the statement of a dis-
covery, but he should also study Faraday for the cultivation of a
scientific spirit, by means of the action and reaction which will
take place between the newly discovered facts as introduced to him
by Faraday and the nascent ideas in his own mind.
It was perhaps for the advantage of science that Faraday, though
thoroughly conscious of the fundamental forms of space, time, and
force, was not a professed mathematician. He was not tempted
to enter into the many interesting researches in pure mathematics
which his discoveries would have suggested if they had been
exhibited in a mathematical form, and he did not feel called upon
either to force his results into a shape acceptable to the mathe-
matical taste of the time, or to express them in a form which
mathematicians might attack. He was thus left at leisure to
do his proper work, to coordinate his ideas with his facts, and to
express them in natural, untechnical language.
It is mainly with the hope of making these ideas the basis of a
mathematical method that I have undertaken this treatise.
529.] We are accustomed to consider the universe as made up of
parts, and mathematicians usually begin by considering a single
particle, and then conceiving its relation to another particle, and so
on. This has generally been supposed the most natural method.
To conceive of a particle, however, requires a process of abstraction,
since all our perceptions are related to extended bodies, so that
the idea of the all that is in our consciousness at a given instant
is perhaps as primitive an idea as that of any individual thing.
Hence there may be a mathematical method in which we proceed
from the whole to the parts instead of from the parts to the whole.
For example, Euclid, in his first book, conceives a line as traced
out by a point, a surface as swept out by a line, and a solid as
generated by a surface. But he also defines a surface as the
230
On the Induction of Electric Currents
boundary of a solid, a line as the edge of a surface, and a point
as the extremity of a line.
In like manner we may conceive the potential of a material
system as a function found by a certain process of integration with
respect to the masses of the bodies in the field, or we may suppose
these masses themselves to have no other mathematical meaning
than the volume-integrals of — V2*, where * is the potential.
In electrical investigations we may use formulae in which the
quantities involved are the distances of certain bodies, and the
electrifications or currents in these bodies, or we may use formulae
which involve other quantities, each of which is continuous through
all space.
The mathematical process employed in the first method is in-
tegration along lines, over surfaces, and throughout finite spaces,
those employed in the second method are partial differential equa-
tions and integrations throughout all space.
The method of Faraday seems to be intimately related to the
second of these modes of treatment. He never considers bodies
as existing with nothing between them but their distance, and
acting on one another according to some function of that distance.
He conceives all space as a field of force, the lines of force being
in general curved, and those due to any body extending from it on
all sides, their directions being modified by the presence of other
bodies He even speaks of the lines of force belonging to a body
as in some sense part of itself, so that in its action on distant
bodies it cannot be said to act where it is not. This, however,
is not a dominant idea with Faraday. I think he would rather
have said that the field of space is full of lines of force, whose
arrangement depends on that of the bodies in the field, and that
the mechanical and electrical action on each body is determined by
the lines which abut on it.
231
The magnetic properties of certain materials and the
electric effects produced by friction were both known
in ancient days. Oersted's experiment with electric
current and a compass showed that electricity and
magnetism are related. Maxwell found the connec-
tion between the two phenomena in his electromag-
netic equations.
15 The Relationship of Electricity and Magnetism
D. K. C. MacDonald
Excerpt from his book, Faraday, Maxwell, and Kelvin, published in 1964.
We know that an electric current can produce forces
on a magnet in its vicinity, or, in other words, an elec-
tric current produces a magnetic "field." Faraday had
shown, moreover, that a changing magnetic field (pro-
duced either by moving a magnet or by varying an elec-
tric current in a coil) could induce an electric current
in a neighboring, but separate, coil of wire. Thus,
through these fundamental experiments of Oersted,
Ampere, and particularly Faraday, various vital facts
had been discovered about how electric currents and
magnets could interact with one another and, as we
have said earlier, these discoveries were already lead-
ing to exciting practical developments such as the elec-
tric telegraph and the submarine cables. But, in broad
terms, what James Clerk Maxwell tried to do was to
build up a more general picture of these interactions
between electric and magnetic effects (or "fields")
233
without worrying so much about actual coils of wire
with electric currents in them, or about how in practice
one actually produced the magnetic fields. Following
Faraday's general lead in concentrating on the "lines
of force" or the "fields," Maxwell tried to work out
directly and quantitatively the interaction in space of
the electric field on the magnetic field, and vice versa,
wherever they might exist. In his mind Maxwell in-
vented, or designed, various semi-mechanical models
to build up his theory, but in the end he could discard
this mental scaffolding and give a complete mathemati-
cal description of electromagnetic behavior which holds
good to this day.
Consider the production of a magnetic field by a cur-
rent of electricity in a coil. We know that such a cur-
rent always involves a movement of electric charge,
so from the electrical point of view we may say that
something is changing all the time. One of the things
Maxwell did was to generalize this discovery boldly,
saying in essence: [I] "A Changing Electric Field Will
Always Produce a Magnetic Field."
But, on the other hand, Faraday had shown that
the movement of a magnet could produce an electric
current, as we have already seen; so on the same lines
this can be generalized to say: [II] "A Changing
Magnetic Field Can Produce an Electric Field"
The ultimate result of James Clerk Maxwell's work
was, in effect, that he expressed these two basic ideas
in precise, quantitative terms, and he came out finally
with what are now known as Maxwell's Equations,
which, as I already have said, remain today the stand-
ard method of predicting how electricity and magnetism
will behave under any given conditions. The acme of
Maxwell's work, however, was his discovery that when
applied in free, empty space his equations took on a
form which is equally descriptive of any undamped
234
The Relationship of Electricity and Magnetism
wave motion propagating itself freely from place to
place. Thus, if you drop a stone into a large pond of
water a ripple or wave will proceed out from that
place, and some of the energy from the falling stone
will radiate outward in the wave from the splash. If
you shout to somebody else some distance away, then it
is a vibration or wave in the air around you which car-
ries the sound to the distant person; or if you rig up a
long, tight rope or string between two points, and then
"twang" the rope, you can see a wave running along
the rope, and this wave carries some of the energy that
you put in the "twang." Again, if there is a violent
storm at sea, the energy from this storm gets carried
over long distances by waves in the ocean; the waves
which smash on the rocks of Newfoundland may well
be getting their energy from a storm a thousand miles
or more out in the Atlantic Ocean. In each of these lat-
ter examples the waves will be damped to some degree
or other. For example, waves traveling on the surface
of the sea lose some energy by dragging deeper layers
of water, by the very fact that water is not entirely free
to move by itself, but has a viscosity or "stickiness,"
which means that the waves ultimately suffer losses by
friction.
The particularly remarkable, and unique, feature of
electromagnetic waves is the fact that they can propa-
gate themselves quite freely without damping through
empty space where no matter whatsoever is present,
but it is not difficult to see from the two italicized state-
ments above that a self-propelled wave motion of the
electromagnetic field might be possible.
Imagine that we have electric and magnetic fields
present in a small region of space, and that the fields
are changing suitably with time. As the electric field
changes at some point in space it will produce a mag-
netic field in the neighborhood, and if things are right
235
this magnetic field will then reinforce the magnetic
field in some regions, and in turn the over-all changing
magnetic field will produce again a fresh electric field
in its neighborhood. What Maxwell's equations showed
was that this process, perhaps somewhat reminiscent of
an endless game of leapfrog, could indeed be self-
maintained, with the energy constantly radiating out-
ward from where the waves started.
But this was not all. Maxwell was able to predict
from this theory, moreover, the speed with which such
an electromagnetic wave should travel in space. This
speed was simply determined by the ratio of two meas-
urements which could be made on electric and mag-
netic quantities in the laboratory, and it turned out
that the speed predicted in this way was very close to
the already known speed of light (about 300,000
km/sec «=* 186,000 miles/sec). Furthermore, it is also
a well-known characteristic of light that it too can
propagate through empty space, as witness the light of
day which reaches us unfailingly from the sun across
about a hundred million miles of empty space. So Max-
well could finally say with confidence that, physically
speaking, light must be a form of electromagnetic
radiation.
Some years after Maxwell's death, Heinrich Hertz
(1857-94) was able to show experimentally, using
electrical apparatus, the direct generation and detection
of the electromagnetic waves predicted by Maxwell.
These "Hertzian waves" are the great-grandfather of
the waves which carry all our radio and television
broadcasts today, and in fact radio waves, television
waves, light waves, X-rays, and gamma rays, are all
members of one and the same family— electromagnetic
waves. In free space they all travel with identically the
same speed, which for convenience we always refer to
as "the velocity of light." What distinguishes one type
236
The Relationship of Electricity and Magnetism
of wave from another is simply its rate of vibration, or
the corresponding wave length (i.e., the distance be-
tween two successive "crests" or "troughs" of a wave).
A typical radio wave vibrates at, or has a frequency
(/) of, about a million times a second (/ = 10G cy-
cles/sec = 1 M c/s), and has a wave length (X) of
about 300 meters. For those who do not mind an equa-
tion, the relationship is very simple, namely f\ = c,
where c denotes, as always in physical science, the
velocity of light. At the other end of the scale, a gamma
ray might have a wave length of only about one ten-
billionth part of a centimeter (\ = 10-10 cm), and a
corresponding frequency of vibration of about three
hundred billion billion cycles/sec (/ = 3 X 1020 c/s).
Electromagnetic Waves
Maxwell's electromagnetic theory also led to intense
discussion later about the fundamental nature of the
electromagnetic waves involved. Many physicists felt
that in order to have a wave at all there had to be
"something" to do the waving or vibrating, and they in-
vented a sort of all-pervading, universal, thin soup or
consomme which they called the "aether." But whether
it is more reasonable to talk about electromagnetic
waves in free space (which still worries some people
for the same sort of reason that "action at a distance"
worried people), or whether it is better to try to think
about an all-permeating, vibrating "aether" is not a
very burning issue today. What matters now is that
Maxwell's Equations are a generally accepted founda-
tion for discussing electromagnetic behavior under the
widest range of possible situations, and also that Max-
well's lead in analyzing electromagnetism by means of
the electric and magnetic fields has led more generally
to the concept of discussing other forms of interaction
237
through some appropriate "field." Indeed, Maxwell
himself was at first very inclined to believe that gravita-
tional attraction must also be propagated in this way,
but he ran up against difficulties with the energy in-
volved which seemed to him then insurmountable.
We have seen that, starting from the picture of "ac-
tion at a distance" between charges of electricity, Max-
well, following Faraday's lead, could reformulate the
problem in terms of a field acting through, and at all
points of, space of which the charged particles are, so
to speak, now just the "terminals" or "end points." The
discovery that this electromagnetic field would vibrate
in free space was a great step toward identifying light
as an electromagnetic wave, since the wave phenome-
non of light (interference, diffraction, etc.) had been
known for a long time. At the same time there had al-
ways been some persistent reasons for regarding light
alternatively as a corpuscular phenomenon, and Ein-
stein was to show, half a century later, that Maxwell's
vibrating electromagnetic aether, when coupled with
Planck's quantum theory first proposed around 1900,
could also then be regarded in a more or less corpuscu-
lar manner. What Planck and Einstein showed was
that the energy in the electromagnetic field could only
exist in certain minimum-sized bundles or "quanta"
dependent in magnitude on the frequency of vibration
and the newly discovered Planck's constant. These
"bundles" of light, or more technically "quanta" of the
electromagnetic field, are generally known today as
photons. So now we can think of electromagnetic in-
teractions as either conveyed by the vibrating aether
or equivalently as conveyed by streams of photons
which will to some extent behave like particles. In deal-
ing with many kinds of interactions, including those
which hold an atomic nucleus together, modern physics
finds it most valuable to be able to think in both these
238
The Relationship of Electricity and Magnetism
terms without being bound to regard one picture as
more necessarily "real" than the other.
239
The formulation of Maxwell's equations opened the
new area of science called electromagnetism, with
far-reaching consequences.
16 The Electromagnetic Field
Albert Einstein and Leopold Infeld
Excerpt from their book entitled the Evolution of Physics published
in 1938 and 1961.
THE REALITY OF THE FIELD
The quantitative, mathematical description of the
laws of the field is summed up in what are called Max-
well's equations. The facts mentioned so far led to the
formulation of these equations but their content is
much richer than we have been able to indicate. Their
simple form conceals a depth revealed only by careful
study.
The formulation of these equations is the most im-
portant event in physics since Newton's time, not only
because of their wealth of content, but also because
they form a pattern for a new type of law.
The characteristic features of Maxwell's equations,
appearing in all other equations of modern physics, are
summarized in one sentence. Maxwell's equations are
laws representing the structure of the field.
Why do Maxwell's equations differ in form and
character from the equations of classical mechanics?
What does it mean that these equations describe the
structure of the field? How is it possible that, from the
results of Oersted's and Faraday's experiments, we can
form a new type of law, which proves so important for
the further development of physics?
241
We have already seen, from Oersted's experiment,
how a magnetic field coils itself around a changing
electric field. We have seen, from Faraday's experi-
ment, how an electric field coils itself around a chang-
ing magnetic field. To outline some of the characteris-
tic features of Maxwell's theory, let us, for the moment,
focus all our attention on one of these experiments,
say, on that of Faraday. We repeat the drawing in
which an electric current is induced by a changing mag-
netic field. We already know that an induced current
appears if the number of lines of force, passing the sur-
face bounded by the wire, changes. Then the current
will appear if the magnetic field changes or the circuit
is deformed or moved: if the number of magnetic lines
passing through the surface is changed, no matter how
this change is caused. To take into account all these
various possibilities, to discuss their particular influ-
ences, would necessarily lead to a very complicated
theory. But can we not simplify our problem? Let us
try to eliminate from our considerations everything
which refers to the shape of the circuit, to its length,
to the surface enclosed by the wire. Let us imagine
that the circuit in our last drawing becomes smaller and
242
The Electromagnetic Field
smaller, shrinking gradually to a very small circuit en-
closing a certain point in space. Then everything con-
cerning shape and size is quite irrelevant. In this limit-
ing process where the closed curve shrinks to a point,
size and shape automatically vanish from our consid-
erations and we obtain laws connecting changes of
magnetic and electric field at an arbitrary point in
space at an arbitrary instant.
Thus, this is one of the principal steps leading to
Maxwell's equations. It is again an idealized experiment
performed in imagination by repeating Faraday's ex-
periment with a circuit shrinking to a point.
We should really call it half a step rather than a
whole one. So far our attention has been focused on
Faraday's experiment. But the other pillar of the field
theory, based on Oersted's experiment, must be consid-
ered just as carefully and in a similar manner. In this
experiment the magnetic lines of force coil themselves
around the current. By shrinking the circular magnetic
lines of force to a point, the second half-step is per-
formed and the whole step yields a connection be-
tween the changes of the magnetic and electric fields
at an arbitrary point in space and at an arbitrary instant.
But still another essential step is necessary. Accord-
ing to Faraday's experiment, there must be a wire test-
ing the existence of the electric field, just as there must
be a magnetic pole, or needle, testing the existence of
a magnetic field in Oersted's experiment. But Maxwell's
new theoretical idea goes beyond these experimental
facts. The electric and magnetic field, or in short, the
electromagnetic field is, in Maxwell's theory, some-
thing real. The electric field is produced by a changing
magnetic field, quite independently, whether or not
243
there is a wire to test its existence; a magnetic field is
produced by a changing electric field, whether or not
there is a magnetic pole to test its existence.
Thus two essential steps led to Maxwell's equations.
The first: in considering Oersted's and Rowland's ex-
periments, the circular line of the magnetic field coil-
ing itself around the current and the changing electric
field, had to be shrunk to a point; in considering.Fara-
day's experiment, the circular line of the electric field
coiling itself around the changing magnetic field had to
be shrunk to a point. The second step consists of the
realization of the field as something real; the electro-
magnetic field once created exists, acts, and changes
according to Maxwell's laws.
Maxwell's equations describe the structure of the
electromagnetic field. All space is the scene of these
laws and not, as for mechanical laws, only points in
which matter or charges are present.
We remember how it was in mechanics. By knowing
the position and velocity of a panicle at one single
instant, by knowing the acting forces, the whole future
path of the particle could be forseen. In Maxwell's
theory, if we know the field at one instant only, we
can deduce from the equations of the theory how the
whole field will change in space and time. Maxwell's
equations enable us to follow the history of the field,
just as the mechanical equations enabled us to follow
the history of material particles.
But there is still one essential difference between me-
chanical laws and Maxwell's laws. A comparison of
Newton's gravitational laws and Maxwell's field laws
244
The Electromagnetic Field
will emphasize some of the characteristic features ex-
pressed by these equations.
With the help of Newton's laws we can deduce the
motion of the earth from the force acting between the
sun and the earth. The laws connect the motion of the
earth with the action of the far-off sun. The earth and
the sun, though so far apart, are both actors in the play
of forces.
In Maxwell's theory there are no material actors.
The mathematical equations of this theory express the
laws governing the electromagnetic field. They do not,
as in Newton's laws, connect two widely separated
events; they do not connect the happenings here with
the conditions there. The field here and novo depends
on the field in the immediate neighborhood at a time
just past. The equations allow us to predict what will
happen a little further in space and a little later in time,
if we know what happens here and now. They allow
us to increase our knowledge of the field by small steps.
We can deduce what happens here from that which
happened far away by the summation of these very
small steps. In Newton's theory, on the contrary, only
big steps connecting distant events are permissible. The
experiments of Oersted and Faraday can be regained
from Maxwell's theory, but only by the summation of
small steps each of which is governed by Maxwell's
equations.
A more thorough mathematical study of Maxwell's
equations shows that new and really unexpected con-
clusions can be drawn and the whole theory submitted
to a test on a much higher level, because the theoretical
consequences are now of a quantitative character and
arc revealed by a whole chain of logical arguments.
245
Let us again imagine an idealized experiment. A small
sphere with an electric charge is forced, by some ex-
ternal influence, to oscillate rapidly and in a rhythmical
way, like a pendulum. With the knowledge we already
have of the changes of the field, how shall we describe
everything that is going on here, in the field language?
The oscillation of the charge produces a changing
electric field. This is always accompanied by a chang-
ing magnetic field. If a wire forming a closed circuit is
placed in the vicinity, then again the changing mag-
netic field will be accompanied by an electric current
in the circuit. All this is merely a repetition of known
facts, but the study of Maxwell's equations gives a
much deeper insight into the problem of the oscillating
electric charge. By mathematical deduction from Max-
well's equations we can detect the character of the
field surrounding an oscillating charge, its structure
near and far from the source and its change with time.
The outcome of such deduction is the electromagnetic
'wave. Energy radiates from the oscillating charge trav-
eling with a definite speed through space; but a trans-
ference of energy, the motion of a state, is character-
istic of all wave phenomena.
Different types of waves have already been consid-
ered. There was the longitudinal wave caused by the
pulsating sphere, where the changes of density were
propagated through the medium. There was the jelly-
like medium in which the transverse wave spread. A
deformation of the jelly, caused by the rotation of the
sphere, moved through the medium. What kind of
changes are now spreading in the case of an electro-
magnetic wave? Just the changes of an electromagnetic
field! Every change of an electric field produces a mag-
246
The Electromagnetic Field
netic field; every change of this magnetic field pro-
duces an electric field; every change of ... , and so
on. As field represents energy, all these changes spread-
ing out in space, with a definite velocity, produce a
wave. The electric and magnetic lines of force always
lie, as deduced from the theory, on planes perpendicu-
lar to the direction of propagation. The wave pro-
duced is, therefore, transverse. The original features of
the picture of the field we formed from Oersted's and
Faraday's experiments are still preserved, but we now
recognize that it has a deeper meaning.
The electromagnetic wave spreads in empty space.
This, again, is a consequence of the theory. If the oscil-
lating charge suddenly ceases to move, then, its field
becomes electrostatic. But the series of waves created
by the oscillation continues to spread. The waves lead
an independent existence and the history of their
changes can be followed just as that of any other ma-
terial object.
We understand that our picture of an electromag-
netic wave, spreading with a certain velocity in space
and changing in time, follows from Maxwell's equa-
tions only because they describe the structure of the
electromagnetic field at any point in space and for any
instant.
There is another very important question. With
what speed does the electromagnetic wave spread in
empty space? The theory, with the support of some
data from simple experiments having nothing to do
with the actual propagation of waves, gives a clear an-
swer: the velocity of an electromagnetic wave is equal
to the velocity of light.
247
Oersted's and Faraday's experiments formed the
basis on which Maxwell's laws were built. All our re-
sults so far have come from a careful study of these
laws, expressed in the field language. The theoretical
discovery of an electromagnetic wave spreading with
the speed of light is one of the greatest achievements in
the history of science.
Experiment has confirmed the prediction of theory.
Fifty years ago, Hertz proved, for the first time, the
existence of electromagnetic waves and confirmed ex-
perimentally that their velocity is equal to that of light.
Nowadays, millions of people demonstrate that elec-
tromagnetic waves are sent and received. Their ap-
paratus is far more complicated than that used by
Hertz and detects the presence of waves thousands of
miles from their sources instead of only a few yards.
248
Instruments borne aloft by artificial satellites and
probes report that our planet is encircled by two zones
containing high-energy radiation against which space
travelers will have to shield themselves.
Radiation Belts Around the Earth
James Van Allen
An article published in Scientific American in 1959.
So far, the most interesting and least
expected result of man's explora-
tion of the immediate vicinity of
the earth is the discovery that our planet
is ringed by a region— to be exact, two re-
gions—of high-energy radiation extend-
ing many thousands of miles into space.
The discovery is of course troubling to
astronauts; somehow the human body
will have to be shielded from this radia-
tion, even on a rapid transit through the
region. But geophysicists, astrophysi-
cists, solar astronomers and cosmic-ray
physicists are enthralled by the fresh im-
plications of these findings. The configu-
ration of the region and the radiation it
contains bespeak a major physical phe-
nomenon involving cosmic rays and solar
corpuscles in the vicinity of the earth.
This enormous reservoir of charged par-
ticles plays a still-unexplained role as
middleman in the interaction of earth
and sun which is reflected in magnetic
storms, in the airglow and in the beauti-
ful displays of the aurora.
The story of the investigation goes
back to 1952 and 1953, before any of
us could think realistically about the use
of earth satellites to explore the environ-
ment of the earth. Parties from our lab-
oratory at the State University of Iowa
spent the summers of those years aboard
Coast Guard and naval vessels, cruising
along a 1,500-mile line from the waters
of Baffin Bay, near the magnetic pole in
the far northwestern corner of Green-
land, southward to the North Atlantic
off the coast of Newfoundland. Along
the way we launched a series of rocket-
carrying balloons— "rockoons." (The bal-
loon lifts a small rocket to an altitude of
12 to 15 miles, whence the rocket car-
ries a modest payload of instruments to
a height of 60 to 70 miles.) Our objec-
tive was to develop a profile of the cos-
mic-ray intensities at high altitudes and
latitudes, and thus to learn the nature of
the low-energy cosmic rays which at
lower altitudes and latitudes are de-
flected by the earth's magnetic field or
absorbed in the atmosphere.
Most of the readings radioed down
from the rockets were in accord with
plausible expectations. Two rockoons
sent aloft in 1953, however, provided us
with a puzzle. Launched near New-
foundland by Melvin Gottlieb and Les-
lie Meredith, they encountered a zone
of radiation beginning at an altitude of
30 miles that was far stronger than we
had expected. At first we were uneasy
about the proper operation of our in-
struments. But critical examination of
the data convinced us that we had un-
questionably encountered something
new in the upper atmosphere.
Significantly these measurements were
made in the northern auroral zone. In
this zone, which forms a ring some 23
degrees south of the north geomagnetic
pole, the incidence of visible auroras
reaches its maximum. Since rockets fired
north and south of the zone had revealed
nothing unusual, we speculated that the
strong radiation played some part in the
aurora. Showers of particles from the
sun, it was thought, come plunging into
the atmosphere along magnetic lines of
force and set off these displays [see "Au-
rora and Airglow," by C. T. Elvey and
Franklin E. Roach; Scientific Ameri-
can, September, 1955]. But the theory
underlying this explanation did not ex-
plain satisfactorily why the aurora and
the high-intensity radiation we had de-
tected should occur in the auroral zone
and not in the vicinity of the geomag-
netic pole itself. Nor could it account
for the high energies required to carry
the solar particles through the atmos-
phere to such relatively low altitudes.
The mystery deepened when we
found in later studies that the radiation
persists almost continuously in the zone
above 30 miles, irrespective of visible
auroral displays and other known high-
altitude disturbances. More discriminat-
ing detectors established that the radia-
tion contains large numbers of electrons.
Our original observations had detected
X-rays only; now it turned out that the
X-rays had been generated by the im-
pact of electrons on the skin of the in-
strument package ( as if it had been the
"target" in an X-ray tube) and on the
sparse atoms of the upper atmosphere
itself. Sydney Chapman and Gordon
Little at the University of Alaska sug-
gested that such a process might well
account for the attenuation of radio sig-
nals in the lower ionosphere of the auro-
ral zones.
T
he International Geophysical Year
gave us our first opportunity to in-
vestigate the "auroral soft radiation" on
a more comprehensive scale. During the
249
STRUCTURE OF RADIATION BELTS revealed by contours of
radiation intensity (black lines) is shown schematically by shading
(left); dots {right) suggest distribution of particles in the two
belts. Contour numbers give counts per second; horizontal scale
summer and fall of 1957 Laurence Ca-
hill and I launched a number of rockoons
off the coast of Greenland and also got
off one successful flight in Antarctica.
The latter flight established that the ra-
diation exists in the southern as well as
the northern auroral zone. In February,
1958, Carl Mclhvain fired a series of
two-stage rockets through visible auro-
ras above Fort Churchill in Canada, and
discovered that the radiation includes
energetic protons (hvdrogen nuclei) as
well as electrons.
Meanwhile all of us had been pushing
a new development that greatly expand-
ed the possibilities for high-altitude re-
search. During the summer of 1955 the
President and other Government author-
ities were finally persuaded that it
might be feasible to place artificial satel-
lites in orbit, and authorized an I. G. Y.
project for this purpose. In January,
1956, a long-standing group of high-
altitude experimentalists, called the
Rocket and Satellite Research Panel,
held a symposium to consider how the
satellites could be most fruitfully em-
ployed. At that meeting our group pro-
posed two projects. One was to put a
satellite into an orbit nearly pole-to-pole
to survey the auroral radiation in both
the north and south auroral zones. Such
orbits, however, did not appear to be
250
Radiation Belts Around the Earth
shows distance in earth radii (about 4,000 miles) from the center
of the earth. Particles in the inner belt may originate with the
radioactive decay of neutrons liberated in the upper atmosphere by
cosmic rays; those in the outer belt probably originate in the sun.
technically feasible in the immediate
future. For the time being we were
forced to abandon the use of a satellite
to probe farther into the auroral soft
radiation. We also suggested that a satel-
lite orbiting over the lower latitudes of
the earth might usefully be employed in
a comprehensive survey of cosmic-ray
intensities over those regions.- This proj-
ect was adopted, and we were author-
ized to prepare suitable experimental
apparatus [see "The Artificial Satellite
as a Research Instrument," by James A.
Van Allen; Scientific American, No-
vember, 1956]. It was planned to place
this apparatus on one of the early Van-
guard vehicles.
The difficulties and failures of the
Vanguard are now history. Sputnik I
stimulated some high government offi-
cials to accept a proposal that a num-
ber of us had been urging for more than
a year: to use the proven Jupiter C
rocket as a satellite-launching vehicle.
As a result on January 31, 1958, Ex-
plorer I went into orbit carrying our
simple cosmic-ray detector and a radio
to broadcast its readings.
In the first reports from stations locat-
ed in the U. S. the intensity of radiation
increased with altitude along the expect-
ed curve. Several weeks later, however,
we began to get tapes from stations in
251
EXPLORER IV AND PIONEER III gave the first detailed picture of the radiation belts.
The Explorer IV satellite (short ellipse) monitored radiation levels for nearly two months
at altitudes up to 1,300 miles. The Pioneer III lunar probe (long ellipse) provided data out
to 65,000 miles. Its orbit is shown distorted because of the earth's rotation during flight.
EXPLORER IV ORBIT covered the entire region 51 degrees north and south of the equator;
the black curve shows a small part of its trace on the earth's surface. More than 25 observa-
tion stations (colored dots) recorded data from several thousand of the satellite's passes.
100.000
\
A
/ w
s
^^^^
10,000 20.000 30,000 40,000
RADIAl DISTANCE FROM CENTER OF EARTH IMIIESI
PIONEER III DATA gave the first confirmation of two distinct rings of particles. Counting
rates on both the outbound (black curve) and the inbound (gray curve) legs of the flight
showed two peaks. The two curves differ because they cover different sections of the belts.
South America and South Africa which
gave us counting rates for much higher
altitudes, due to the eccentricity of the
satellite's orbit. These records brought us
a new surprise. At high altitudes over the
equatorial region the apparent counting
rate was very low; in some passes it
dropped to zero for several minutes. Yet
at lower altitudes the rate had quite
"reasonable" values— from 30 to 50
counts a second. Again we were uneasy
about the trustworthiness of the instru-
ments. The only alternative seemed to
be that cosmic rays do not strike the
uppermost layers of the atmosphere over
the tropics, and we were quite unable
to accept this conclusion.
Our uneasiness was increased bv the
incompleteness of our early data. The
Explorer I apparatus broadcast its obser-
vations continuously, but its signals
could be picked up only intermittently,
when the satellite came within range of
a ground station. Our original apparatus,
designed and developed by George Lud-
wig for the Vanguard satellites, included
a magnetic-tape recorder which could
store its observations for a complete orbit
around the earth and then report them in
a "burst" on radio command from the
ground.
T)y early February, working with the
*-* Jet Propulsion Laboratory, we had
convertea this apparatus for use in the
Explorer II satellite. The first attempt to
get it into orbit failed. A second rocket
placed Explorer III, carrying identical
apparatus, in orbit on March 26. This
satellite fully confirmed the anomalous
results of Explorer I. At altitudes of 200
to 300 miles the counting rate was low.
When the satellite went out to 500 to
600 miles, the apparent rate ascended
rapidly and then dropped almost to zero.
One dav, as we were puzzling over the
first tapes from Explorer III, Mcllwain
suggested the first plausible explanation
for their peculiar readings. He had just
been calibrating his rocket instruments,
and called our attention to something
that we all knew but had temporarily
forgotten: A sufficiently high level of
radiation can jam the counter and send
the apparent counting rate to zero. We
had discovered an enormously high level
of radiation, not a lack of it. As Ernest
Ray, a member of our group, inaccu-
rately but graphically exclaimed: "Space
is radioactive!"
During the next two months Explorer
III produced a large number of playback
records, every one of which showed the
same effect. At low altitudes the count-
ing rate was reasonably attributable to
252
Radiation Belts Around the Earth
DETECTOR FOR CHANNEL 4 DETECTORS FOR CHANNELS I AND 3
U-_-^
ELECTRONICS, BATTERIES. HIGH POWER TRANSMITTER DETECTOR FOR CHANNELS 2 AND 5
PLASTIC
SCINTILLATOR
PHOTOMULTIPL1ER
TUBE
CESIUM IODIDE
SCINTILLATOR
PHOTOMULTIPLIER
TUBE
GEIGER TUBE
SHIELDED
GEIGER TUBE
LINEAR
AMPLIFIER
Ami
NON-LINEAR
NETWORK
1 2 3 4 2048
JUUUUL
1 2 3 4 64
JUUUUL
1 2 3 ...16
DISCRIMINATOR
CALIBRATION RELAY
-| SCALE OF 2048 \-
-| SCALE OF 64 [-
SCALE
OF 16
1 2 3. ..128
JUUUl
SCALE
OF 128
ELECTROMETER
LP
J"
nnnfir^n J1
LOW POWER
TRANSMITTER
lT
-\ CHANNEL2
CHANNEL 3
v-
HIGH POWER
TRANSMITTER
<
<
ruiii
CHANNEL 1
~*«-
CHANNEL 2
J-Ju
CHANNEL 3
ir
f
***t
,.\in ■>- . . J»f
< 64 COUNTS PER CYCLE >
LJ1
1}
_r
4F*
•2048 COUNTS PER CYCLE-
UlflLIl
w
w
ft
w
r
trr
ill
CHANNEL 5
I i I i
EXPLORER IV INSTRUMENTS were designed to give a detailed rather than individual particles. Shielded and ""shielded Geiger
LXTLUKt-K IV inaiKUJncn 19 "....* Plastic scin. mbes could be compared to estimate the penetrability of the radia-
Dicture of the nature and intensity of the radiation, riasuc sun luuoa r
picture o. me nature a a » energies; two tion. Radio signals suggested by the red curves in upper drawing
tillator counted only charged particles above certain energies, •» , J , . . ,; „ „„j i0,„r nUv*A thrnnsh a
different scaling factors adapted i, to both high and low counting were recorded by S">undj.a -s and a ler pi ,e ihroug •
rates. Cesium-iodide scintillator measured the total energy input multichannel oscillograph to y.eld record, hke that shown below.
253
MILES
1,500 1,000 5C0
MUES
500 1,000 1,500
TWO SETS OF CONTOURS from readings on opposite sides of
the earth Hell and center) show the northern and southern "horns"
of radiation, which point toward the auroral zones; the contour
numbers show radiation intensity in counts per second. The "tipped"
cosmic rays. At higher altitudes— the pre-
cise height depended on both latitude
and longitude— the count increased to
very high values. Up to the points at
which the counter jammed, it showed
counting rates more than 1,000 times
the theoretical expectation for cosmic
rays. From the rate of increase and the
length of the periods of jamming we
judged that the maximum count prob-
ably went to several times this level.
Since the radiation appeared to resem-
ble the auroral soft radiation, we would
not have been surprised to find it in the
auroral zone or along the magnetic lines
of force that connect these zones. But in
the equatorial latitudes these lines of
force lie much farther out in space than
the altitudes attained by the satellites.
On May 1 of last year we were able
to report with confidence to the National
Academy of Sciences and the American
Physical Society that Explorers I and
III had discovered a major new phenom-
enon: a very great intensity of radiation
above altitudes of some 500 miles over
the entire region of their traverse, some
34 degrees north and south of the equa-
tor. At the same time we advanced the
idea that the radiation consists of
charged particles— presumably protons
and electrons— trapped in the magnetic
field of the earth.
We could rule out uncharged particles
and gamma and X-rays because they
would not be confined by the magnetic
field, and so would be observed at lower
altitudes. The possibility that the earth's
magnetic field might act as a trap for
charged particles was first suggested by
the Norwegian physicist Carl Stormer
in a classical series of papers beginning
some 50 years ago, and there was a
considerable body of evidence for the
existence of low-energy charged parti-
cles throughout our solar system and
specifically in the vicinity of the earth.
But there had been no indication that
these particles would possess the high
energies we had detected.
From Stormer's theoretical discussion
and our own observations we evolved a
rough picture of the trapping mechan-
ism. When a fast-moving charged parti-
cle is injected into the earth's magnetic
field, it describes a corkscrew-shaped
trajectory, the center line of which lies
along a magnetic line of force. The turns
of the helical path are quite open over
the equator but become tighter as the
particle reaches the stronger magnetic
field toward the poles [see illustration at
bottom of opposite page]. At the lower
end of its trajectory the particle goes into
a flat spiral and then winds back along
a similar path to the other hemisphere,
making the transit from one hemisphere
to the other in a second or so. During
this time its line of travel shifts slightly,
SO that the particle drifts slowly around
the earth as it corkscrews from hemi-
sphere to hemisphere. An electron drifts
from west to east; a proton, in the op-
posite direction. At each end of its path
the particle descends into regions of
higher atmospheric density; collisions
with the atoms of atmospheric gases
cause it gradually to change its trajectory
and to lose energy. After a period of days
or weeks the particle is lost into the lower
atmosphere.
r 1 1here was obviously an urgent scien-
-*- tific need to extend these observa-
tions with equipment of greater dynamic
range and discrimination. In April of
1958 we persuaded several Federal
agencies to support further satellite
flights of our radiation equipment as an
adjunct to the I. G. Y. program, and we
received the enthusiastic support of the
National Academy of Sciences for the
continuation of our work. We also per-
suaded the Army Ballistic Missile Agen-
cy and the Cape Canaveral Air Force
Base to try to place the satellite in an
orbit more steeply inclined to the equa-
tor; at an inclination of about 50 degrees
to the equator it would cover a much
greater area of earth and skim the edges
of both auroral zones.
Working night and dav, we set out at
once to build new apparatus of a more
discriminating nature. We retained the
Geiger tube, which we had used in pre-
vious satellites, as a basic "simple-mind-
ed" detector. To be ready for the highest
intensities of radiation, however, we
used a much smaller tube that would
yield a lower count in a given flux of
radiation, and we hooked it into a circuit
that would scale down its count by a
much larger factor. To obtain a better
idea of the penetrability of the radiation
254
Radiation Belts Around the Earth
MAGNETIC AXIS
MAGNETIC EQUATOR
drawing al right shows the essential symmetry
axis. The structure of the radiation zone was
we shielded a similar Geiger tube with a
millimeter of lead. As a more discriminat-
ing particle detector we adopted a plas-
tic scintillator and photomultiplier tube
to respond to electrons with an energy
of more than 650,000 electron volts and
to protons of more than 10 million elec-
tron volts. Finally we glued a thin cesi-
um-iodide crystal to the window of an-
other photomultiplier tube; the light
emitted by the crystal when it was ir-
radiated would measure the over-all in-
put of energy rather than the arrival of
individual particles. To keep out light
when the crystal faced the sun, we
shielded it with thin, opaque nickel foil.
A special amplifier gave this detector a
large dynamic range extending from
about .1 erg per second to 100,000 ergs
per second.
Explorer IV carried this apparatus in-
to orbit on July 26, and sent down data
for almost two months. Magnetic tapes
from some 25 observing stations flowed
in steadily from late July to late Septem-
ber; altogether we obtained some 3,600
recorded passes of the satellite. A typical
pass was readable for several minutes;
some of the best were readable for up to
20 minutes, a large fraction of the time
required for the satellite to make a turn
around the earth. We are still analyzing
this mass of data, but the preliminary
results have already proved to be en-
lightening.
The readings have confirmed our ear-
lier estimates of the maximum levels of
radiation. Moreover, we have extended
of the radiation around the earlh*s magnetic
built up from hundreds of observed points.
our observations to more than 50 degrees
north and south of the equator and have
been able to plot the intensity of the
radiation at various latitudes and longi-
tudes for altitudes up to 1,300 miles.
The intensity contours follow the shape
of the earth in the equatorial region, but
as they approach high northern and
southern latitudes they swing outward,
then inward and sharplv outward again
to form "horns" reaching down toward
the earth near the auroral zones [see
illustrations at the top oj these two
pages]. The entire picture so far is com-
pletely consistent with the magnetic-
trapping theory.
It was clear from the contours that
Explorers I, III and IV penetrated onlv
the lower portion of the radiation belt.
As early as last spring we began to make
hypothetical extensions of the observed
contours out to a distance of several
thousand miles. One of these speculative
diagrams showed a single, doughnut-
shaped belt of radiation with a ridge
around the northern and southern edges
of its inner circumference, correspond-
ing to the horns of the contours. Another
showed two belts— an outer region with
a banana-shaped cross section that ex-
tended from the northern to the southern
auroral zone and an inner belt over the
equator with a bean-shaped cross section
[see illustration on pages 40 and 41].
The latter diagram seemed to fit the con-
tours better. In our seminars and after-
hour discussions Mcllwain held out for
the two-belt theory. The rest of us tend-
ed to agree with him but preferred to
stay with the single "doughnut" because
of its simplicity.
T^o take the question out of the realm
■*- of speculation we had to secure
measurements through the entire region
of radiation. In May, therefore, I ar-
ranged to have one of our radiation de-
tectors carried aboard the lunar probes
planned for the fall of 1958. On October
mm&**
TRAPPED PARTICLES spiral rapidly back and forth along a corkscrew shaped path
whose center is a magnetic line of force. At the same time they drift slowly around the earth
{broken arrows). Electrons I negative) and protons I positive* drift in opposite directions.
255
11, 12 and 13 Pioneer I, the first lunar
probe, carried our instruments nearly
70,000 miles out from the earth. Though
its readings were spotty, they confirmed
our belief that the radiation extended
outward for many thousands of miles,
with its maximum intensity no more than
10,000 miles above the earth.
The next attempted moon shot, Pio-
neer II, was a fizzle. Pioneer III, how-
ever, went off beautifully on December
6. Although this rocket was intended to
reach the vicinity of the moon, we were
almost as pleased when it failed to do
so, for it gave us excellent data on both
the upward and downward legs of its
flight, cutting through the radiation re-
gion for 65,000 miles in two places.
The observations on both legs showed
a double peak in intensity [see illustra-
tion at bottom of page 42], establishing
that there are indeed two belts rather
than one. The inner belt reaches its
peak at about 2,000 miles from the earth,
the outer one at about 10,000 miles.
Bevond 10,000 miles the radiation in-
tensity diminishes steadily; it disappears
almost completely beyond 40,000 miles.
The maximum intensity of radiation in
each belt is about 25,000 counts per sec-
ond, equivalent to some 40,000 parti-
cles per square centimeter per second.
Most of us believe that this great
reservoir of particles originates largely
in the sun. The particles are somehow
injected into the earth's magnetic field,
where they are deflected into corkscrew
trajectories around lines of force and
trapped. In this theoretical scheme the
radiation belts resemble a sort of leaky
bucket, constantly refilled from the sun
and draining away into the atmosphere.
A particularly large influx of solar par-
ticles causes the bucket to "slop over,"
mainly in the auroral zone, generating
visible auroras, magnetic storms and re-
lated disturbances. The normal leakage
may be responsible for the airglow which
faintly illuminates the night sky and may
also account for some of the unexplained
high temperatures which have been ob-
served in the upper atmosphere.
This solar-origin theory, while attrac-
tive, presents two problems, neither of
which is yet solved. In the first place
the energy of many of the particles we
have observed is far greater than the pre-
sumed energy of solar corpuscles. The
kinetic energy of solar corpuscles has
not been measured directly, but the
time-lag between a solar outburst and
the consequent magnetic disturbances
HEAD OF EXPLORER IV includes nose rone (left), instrument "payload" (center) and
protective shell i right). Payload includes four detectors, two radio transmitters, batteries
and associated electronic circuity. The outer shell is approximately six inches in diameter.
on earth indicates that the particles are
slow-moving and thus of relatively low
energy. It may be that the earth's mag-
netic field traps only a high-energy frac-
tion of the particles. Alternatively, some
unknown magnetohydrodynamic effect
of the earth's field may accelerate the
sluggish particles to higher velocities.
Some such process in our galaxv has
been suggested as responsible for the
great energies of cosmic rays. The second
problem in the solar-origin theory is that
it is difficult to explain how charged
particles can get into the earth's mag-
netic field in the first place. We believe
that neither problem is unsolvable.
Nicholas Christofilos of the University
of California and the Soviet physicist
S. N. Vernov have suggested an entirelv
different theory of how the radiation
originates. They note that neutrons are
released in large numbers in the earth's
upper atmosphere by the impact of cos-
mic rays. These neutrons, being un-
charged, can travel through the mag-
netic field without deflection. In due
course some of them decav there into
electrons and protons, which are trapped.
Our group agrees that particle-injec-
tion of this sort is going on, and at a rate
which can be easily calculated; but we
feel for a number of reasons that it can-
not be the main source of radiation-belt
particles. If we are right in supposing
that the radiation belts provide the "res-
ervoir" for the aurora, the neutron hy-
pothesis cannot account for more than
one 10,000th of the auroral energy out-
put. Even if the association between
the radiation belts and the aurora turns
out to be fortuitous, preliminary indica-
tions both from our work and from the
Russian experience with Sputnik III
suggest that most of the particles in the
radiation belt have much lower energies
than those of particles that would be
produced by neutron decay. A full
knowledge of the energy distribution of
the particles will aid greatly in clarifying
their origin.
Neither theory explains why there
should be two belts rather than one. It is
tempting to combine the two theories
and suppose that the inner belt orig-
inates with "internal injection"— i.e., neu-
tron-decav products— and the outer one
with "external injection" of solar cor-
puscles. The two-belt configuration may
of course be a transitory phenomenon,
though the data from Explorer IV and
Pioneer III indicate that the separate
belts persisted in essentially the same
form for at least five months. We should
bear in mind, however, that 1958 was
a \ ear of great solar activity. Three years
256
Radiation Belts Around the Earth
FOUR-STAGE ROCKET launched the Pioneer III moon probe on
December 6, 1958. Though the flight failed to reach the moon, its
outbound leg gave a continuous record of radiation out to 65,000
miles; the inbound leg gave data between 30,000 and 10,000 miles.
from now we may well find a much
lower over-all intensity and perhaps a
different structure altogether.
In addition to these possible long-term
changes, there may be short-term fluc-
tuations in the belts. While we feel sure
that the influx and leakage of particles
must balance in the long run, a major
solar outbreak may temporarily increase
the intensity of the radiation many-fold.
If we were to detect such fluctuations
and were to find that they coincide with
solar outbursts on the one hand and
with terrestrial magnetic disturbances
on the other, we would have a plain
lead to the origin of the particles. Be-
fore long we hope to launch a satellite
that will monitor radiation levels for
at least a year.
/"\ur measurements show that the max-
^-^ imum radiation level as of 1958 is
equivalent to between 10 and 100 roent-
gens per hour, depending on the still-
undetermined proportion of protons to
electrons. Since a human being exposed
for two days to even 10 roentgens would
have onlv an even chance of survival, the
radiation belts obviously present an ob-
stacle to space flight. Unless some prac-
tical way can be found to shield space-
travelers against the effects of the radia-
tion, manned space rockets can best take
off through the radiation-free zone over
the poles. A "space station" must orbit
below 400 miles or beyond 30,000 miles
from the earth. We are now planning a
satellite flight that will test the efficacy
of various methods of shielding.
The hazard to space-travelers may not
end even when they have passed the
terrestrial radiation belts. According to
present knowledge the other planets of
our solar system mav have magnetic
fields comparable to the earth's and thus
may possess radiation belts of their own.
The moon, however, probablv has no
belt, because its magnetic field appears
to be feeble. Lunar probes should give
us more definite information on this
point before long.
257
How does the brain work? Part of the answer lies in
electrophysiology, the study of the relation between
electricity and nervous stimulation.
18 A Mirror for the Brain
W. Grey Walter
A chapter from his book The Living Brain published in 1963.
The Greeks had no word for it. To them the
brain was merely "the thing in the head," and completely
negligible. Concerned as so many of them were about man's
possession of a mind, a soul, a spiritual endowment of the
gods, it is strange they did not anticipate our much less enter-
prising philosophers of some score of centuries later, and in-
vent at least a pocket in the head, a sensorium, to contain it.
But no, the Greeks, seeking a habitation for the mind, could
find no better place for it than the midriff, whose rhythmic
movements seemed so closely linked with what went on in
the mind.
The Hebrews also attributed special dignity to that part of
the body; thence Jehovah plucked man's other self. Old ideas
are not always as wide of the mark as they seem. The rhythm
of breathing is closely related to mental states. The Greek
word for diaphragm, phren, appears in such everyday words
as frenzy and frantic, as well as in the discredited phrenology
and the erudite schizophrenia.
259
Above the midriff the classical philosophers found the
vapours of the mind; below it, the humours of the feelings.
Some of these ideas persisted in physiological thought until
the last century and survive in the common speech of today.
Hysteric refers by derivation to the womb. The four basic
human temperaments were: choleric, referring to the gall
bladder; phlegmatic, related to inflammation; melancholic,
black bile; and sanguine, from the blood. This classification
of temperaments was revived by a modern physiologist, Pav-
lov, to systematize his observations of learning.
As in nearly all notions that survive as long as these fossils
of language have survived, there is an element of truth, of
observation, in them. States of mind are certainly related to
the organs and liquors designated, and may even be said in
a sense to originate in them. The philosopher, William James,
was responsible with Lange for a complete theory of emotion
which invoked activity in the viscera as the essential precursor
of deep feeling. Some of the most primitive and finest phrases
in English imply this dependence of sincere or deep emotion
on heart or bowels. But communication of thought is so rapid
that the Greeks overlooked the existence or need of a relay
station. And no doubt it is for the same reason that we all
seem particularly given to the same error of over-simplification
when we first begin, or refuse to begin, to consider how the
mind works. We know what makes us happy or unhappy.
Who, in the throes of sea-sickness, would think of dragging
in the brain to account for his melancholy state?
More curious still is Greek negligence of the brain, con-
sidering their famous oracular behest, "Know thyself." Here
indeed was speculation, the demand for a mirror, insistence
upon a mirror. But for whom, for what? Was there, among
260
A Mirror for the Brain
the mysteries behind the altar, concealed perhaps in the
Minerva myth, a suspicion of something more in the head
than a thing, and that the organ which had to do the knowing
of itself must be an organ of reflection?
The brain remained for more than two thousand years in
the dark after its coming of age. When it was discovered by
the anatomist, he explored it as a substance in which might
be found the secret dwelling of intelligence; for by that time
the mind had moved from the diaphragm to the upper story,
and Shakespeare had written of the brain, "which some sup-
pose the soul's frail dwelling-house." Dissection was high
adventure in those days. Most people believed what an
ironical writer today was "astonished to leam," that "it is
possible for anger, envy, hatred, malice, jealousy, fear and
pride, to be confined in the same highly perishable form of
matter with life, intelligence, honesty, charity, patience and
truth.*' The search for such prize packets of evil and virtue
in the brain tissue, dead or alive, could only lead to disap-
pointment. The anatomist had to be satisfied with weighing
the "grey matter" — about 50 ounces for man and 5 less for
woman — and making sketches of the very complicated and
indeed perishable organisation of nerves and cells which his
knife revealed. He could do little more. It should enlighten
us at once as to the essential character of brain activity, that
there was no possible understanding of ihe mechanism of
the brain until the key to it, the electrical key, was in our
hands.
There were some flashes of foresight, sparks in the scien-
tific dark, before Galvani put his hand on the key. What gen-
erated all the speculations of the day was a new notion in
261
science, the conception of physical motion which began to
acquire importance with Galileo and continued with Newton
and into our own times with Rutherford and Einstein. First
among these imaginative flashes may be mentioned the novel
proposal made by the 16th Century philosopher, Hobbes,
when disputing the dualist theory of Descartes. The French
philosopher contemplated a non-spatial mind influencing the
body through the brain, and suggested the pineal gland as
the rendezvous for mind and matter. The proposal advanced
by Hobbes, in rejecting this popular theory, was that thought
should be regarded as being produced by bodies in motion.
Hobbes was born in the year of the Spanish Armada; the
Royal Society had received its charter seventeen years before
he died in 1679.
The controversy about ihe residential status of the mind
is almost as much out of date as that in which the non-
existence of motion seemed to be proved by the hare and
tortoise fable. But the value of Hobbes' speculation was en-
during; the observation and correlation of mental and phys-
ical phenomena are today a routine of physiological research.
More specific than the speculation of Hobbes was that of
Dr. David Hartley about a century later. Hartley in 1749
anticipated by two hundred years the kind of theory of mental
function for which evidence has been found in the last year
or two. His "Observations on Man, his Frame, his Duty and
his Expectations" is a milestone in the history of English
thought. Hartley, a contemporary of Newton and Hume, was
a pioneer of what he termed the "doctrine of mechanism."
According to this, he suggested, mental phenomena are de-
rived from rhythmic movements in the brain — vibrations, he
called them; upon these is superimposed a fine structure of
262
A Mirror for the Brain
"vibratiuncles" which give thought and personality their
subtle shades and variations. Hartley realised quite well the
value of the plastic and compact virtues such a system might
have. He was also the first to develop the theory of "associa-
tion of ideas" in a rigorous form, relating this to his "vibra-
tiuncles" in a manner which we should now consider strictly
scientific in the sense that it is susceptible to experimental
test. It is difficult for us to appreciate the originality of his
notions, the gist of which is now a commonplace of electro-
physiology.
Hartley wrote nearly half a century before Galvani ( 1737-
1798 ) and with him we might say farewell to fancy. But to
pass over the famous Galvani- Volta controversy with the
bald statement that the one claimed to have discovered elec-
tricity in animals and the other its generation by metals, would
be unfair to any reader who may not know how strangely
truth came out of that maze of error.
The incident began with an experiment made by Luigi and
Lucia Galvani in the course of their long and patient study
of that still fresh mystery, electricity. The word had been in
use since William Gilbert coined it in the 16th century from
elektron, meaning amber, another pretty semantic shift; and
Henry Cavendish had already, eight years before the inci-
dent, determined the identity of its dynamic laws with those
of gravitation. Everybody in high society was familiar with
the effects of discharges from Leyden jars upon the lifeless
muscles of executed criminals; and Louis XV had, in the
words of Silvanus Thomson, "caused an electric shock from a
battery of Leyden jars to be administered to 700 Carthusian
monks joined hand to hand, with prodigious effect." But in
Bologna in 1790 the professor of anatomy had a notion that
263
it was atmospheric electricity which acted upon the muscle
tissues of animals. On a stormy evening, one version of the
story goes, he and his wife had the curious idea of testing
this point by tying a dead frog to the top of the iron balustrade
of the court-yard, apparently using copper wire to hold it
by the leg. They expected that, as the storm approached, the
frog would be convulsed by electric shocks. And, as they
watched the thunder cloud come near, so indeed it happened;
the dead frog, hanging against the iron grill, twitched in re-
peated convulsions.
Further experiments convinced the Galvani that they had
witnessed a form of electricity derived from living processes,
not merely from the atmosphere. He published a famous ac-
count of his experiments on the relation of animal tissue to
electricity: De viribus Electricitatis in Motu Musculari Com-
mentarius (1791). Volta seized upon this to refute the whole
of Galvani's thesis, repeating his experiment not only with-
out the storm but without the frog, proving that the elec-
tricity in question could be generated by copper and zinc
sheets. This "current electricity" as it was called, was there-
fore metallic, and no nonsense about any animal variety. So
ended a controversy and a friendship. So began the science
of electrical engineering.
Eppur, the Galvani might have repeated, si muove. For
their discredited experiment had truly revealed, not indeed
what they supposed, but something more wonderful. What
had happened was that, swaying in the wind, the suspended
frog had come into contact with the iron bars, between which
and the copper wire a current had been generated, activating
its muscles. The Galvani had demonstrated the electrical
aspect of nervous stimulation.
264
A Mirror for the Brain
This was an event as important to the physiologist as its
counter-event was to the physicist; it was the starting-point
of that branch of the science with which we are concerned
here, electrophysiology.
Volta's counter-demonstration led directly to the invention
of the electric battery, and economic opportunity evoked
electrical engineering from the Voltaic pile. There was no
such incentive for research when, a generation later, the
existence of animal electricity was proved. Instead, the dis-
covery was exploited by the academic dilettante and the
quack. The Aristotelian doctors of the period, assuming that
where there is electricity there is magnetism, saw in it proof
also of Mesmer's "Propositions" which had been published
in his "Memoire sur la Decouverte du Magnetisme Animal"
in 1779, floundering deeper into mystification than Dr. Mes-
mer himself, who had at least declared in his "Memoire" that
he used the term analogically, and that he "made no further
use of electricity or the magnet from 1776 onwards."
There is still controversy about the origin and nature of
animal electricity. Nobody who has handled an electric eel
will question the ability of an animal to generate a formidable
voltage; and the current is demonstrably similar in effect to
that of a mineral dry cell. On the other hand, there is no evi-
dence that the electric energy in nerve cells is generated by
electro-magnetic induction or by the accumulation of static
charge. The bio-chemist finds a complicated substance, acetyl-
choline, associated with electric changes; it would be reason-
able to anticipate the presence of some such substance having
a role at least as important as that of the chemicals in a Le-
clanche cell.
We know that living tissue has the capacity to concentrate
265
potassium and distinguish it from sodium, and that neural
electricity results from the differential permeability of an
inter-face, or cell-partition, to these elements, the inside of
a cell being negatively charged, the outside positively.
Whether we call this a chemical or an electrical phenomenon
is rather beside the point. There would be little profit in argu-
ing whether a flash-lamp is an electrical or chemical device;
it is more electrical than an oil lamp, more chemical than a
lightning flash. We shall frequently refer to changes of po-
tential as electrical rhythms, cycles of polar changes, more
explicitly electro-chemical changes. We shall be near the truth
if we keep in mind that electrical changes in living tissue,
the phenomena of animal electricity, are signs of chemical
events, and that there is no way of distinguishing one from
the other in the animal cell or in the mineral cell. The current
of a nerve impulse is a sort of electro-chemical smoke-ring
about two inches long travelling along the nerve at a speed
of as much as 300 feet per second.
The neglect and mystification which obscured Galvani's
discovery, more sterile than any controversy, forced electro-
physiology into an academic backwater for some decades. A
few experiments were made; for example, by Biedermann,
who published a 2-volume treatise called Electrophysiology,
and by Dubois-Reymond, who introduced Michael Fara-
day's induction coil into the physiological laboratory and the
term faradisation as an alternative to galvanisation into the
physiotherapist's vocabulary. Faraday's electrical and elec-
trifying research began in 1831, the date also of the foun-
dation of the British Association for the Advancement of
Science; but physiology long remained a backward child of
the family.
266
A Mirror for the Brain
Hampered though these experimenters were by lack of
trustworthy equipment — they had to construct their own
galvanometers from first principles — they gradually accumu-
lated enough facts to show that all living tissue is sensitive
in some degree to electric currents and, what is perhaps more
important, all living tissue generates small voltages which
change dramatically when the tissue is injured or becomes
active.
These experiments were not concerned with the brain; they
were made on frog's legs, fish eggs, electric eels and flayed
vermin. Nor could the brain be explored in that way.
Following life through creatures you dissect,
You lose it in the moment you detect
It took a war to bring the opportunity of devising a tech-
nique for exploring the human brain — and two more wars to
perfect it Two medical officers of the Prussian army, wan-
dering through the stricken field of Sedan, had the brilliant if
ghoulish notion to test the effect of the Galvanic current on
the exposed brains of some of the casualties. These pioneers
of 1870, Fritsch and Hitzig, found that when certain areas at
the side of the brain were stimulated by the current, move-
ments took place in the opposite side of the body.
That the brain itself produces electric currents was the dis-
covery of an English physician, R. Caton, in 1875.
This growing nucleus of knowledge was elaborated and
carried further by Ferrier in experiments with the "Faradic
current.'* Toward the end of the century there was a spate of
information which suggested that the brain of animals pos-
sessed electrical properties related to those found in nerve and
muscle. Prawdwicz-Neminski in 1913 produced what he
267
called the "electro-cerebrogranT of a dog, and was the first to
attempt to classify such observations.
The electrical changes in the brain, however, are minute.
The experiments of all these workers were made on the ex-
posed brains of animals. There were no means of amplifica-
tion in those days, whereby the impulses reaching the exterior
of the cranium could be observed or recorded, even if their
presence had been suspected. On the other hand, the grosser
electrical currents generated by the rhythmically contracting
muscles of the heart were perceptible without amplification.
Electro-cardiography became a routine clinical aid a gen-
eration before the invention of the thermionic tube made it
possible to study the electrical activity of the intact human
brain.
From an unexpected quarter, at the turn of the century,
came an entirely new development. Turn up the section on
the brain in a pre-war textbook of physiology and you will
find gleanings from clinical neuro-anatomy and — Pavlov. Al-
most as if recapitulating the history of physiological ideas,
Pavlov's work began below the midriff. He found that the
process of digestion could not be understood without refer-
ence to the nervous system, and so commenced his laborious
study of learning in animals.
In the gospel according to Stalin, Pavlov founded not
merely a branch of physiology as Galvani had done, but a
whole new science — Soviet physiology. His work indeed was
original; it owed nothing to Galvani, lying quite outside elec-
trophysiology, to which it was nevertheless eventually, though
not in Pavlov's day, to contribute so much in the way of under-
standing.
A Mirror for the Brain
For nearly two generations Pavlov's experiments were the
major source of information on brain physiology. Workers in
the English laboratories had not permitted themselves to ex-
plore further than the top of the spinal cord. One took an
anatomical glance at the brain, and turned away in despair.
This was not accountable to any peculiar weakness of physio-
logical tradition but to the exigencies of scientific method
itself. A discipline had been building up through the cen-
turies which demanded that in any experiment there should
be only one variable and its variations should be measurable
against a controlled background. In physiology this meant
that in any experiment there should be only one thing at a
time under investigation — one single function, say, of an
organ — and that the changes of material or function should
be measurable. There seemed to be no possibility of isolating
one single variable, one single mode of activity, among the
myriad functions of the brain. Thus there was something like
a taboo against the study of the brain. The success of Pavlov
in breaking this taboo early in the century was due to his
contrivance for isolating his experimental animals from all
but two stimuli; his fame rests on his measurement of re-
sponses to the stimuli.
There was no easy way through the academic undergrowth
of traditional electrophysiology to the electrical mechanisms
underlying brain functions. The Cambridge school of electro-
physiology, under a succession of dexterous and original ex-
perimenters beginning toward the end of the last century,
developed its own techniques in special fields of research, par-
ticularly in the electrical signs of activity in muscles, nerves
and sense organs. At the same time, the Oxford school under
the leadership of Sherrington was beginning to unravel some
269
of the problems of reflex function of the spinal cord. In both
these schools the procedure adopted, to comply with the
traditional requirements of scientific method, was to dissect
out or isolate the organ or part of an organ to be studied. This
was often carried to the extreme of isolating a single nerve
fibre only a few thousandths of a millimetre in diameter, so
as to eliminate all but a single functional unit.
Imagine, then, how refreshing and tantalizing were the
reports from Pavlov's laboratory in Leningrad to those en-
gaged on the meticulous dissection of invisible nerve tendrils
and the analysis of the impulses which we induced them to
transmit After four years spent working literally in a cage
and chained by the ankle — not for punishment but for elec-
trical screening — enlargement came when my professor of
that date, the late Sir Joseph Barcroft, assigned me to estab-
lishing a laboratory in association with a visiting pupil of
Pavlov, Rosenthal. We spent a year or so on mastering the
technique and improving it by the introduction of certain
electronic devices. The Russian results were confirmed. To
do more than this would have required staff and equipment
far beyond the resources of the Cambridge laboratory.
Meanwhile, another major event in the history of physiology
had taken place. Berger, in 1928, at last brought Hartley's
vibrations into the laboratory and with them a method which
seemed to hold out the promise of an investigation of elec-
trical brain activity as precise as were the reflex measure-
ments of Pavlov. When Pavlov visited England some time
after we heard of this, as the English exponent of his work
I had the privilege of discussing it with him on familiar terms.
Among other things, I asked him if he saw any relation be-
tween the two methods of observing cerebral activity, his
270
A Mirror for the Brain
method and Berger's. The latter, I was even then beginning
to suspect, might in some way provide a clue to how the con-
ditioning of a reflex was effected in the brain. But Pavlov
showed no desire to look behind the scenes. He was not in
the least interested in the mechanism of cerebral events; they
just happened, and it was the happening and its consequence
that interested him, not how they happened. Soviet physiology
embalmed the body of this limited doctrine as mystically as
the body of Lenin, for the foundations of their science. The
process of conditioning reflexes has a specious affinity with
the Marxian syllogism. Others have found in the phenomena
sufficient substantiation for a gospel of Behaviourism.
Pavlov was before his time. He would have been a greater
man, his work would have been more fertile in his lifetime,
and Russian science might have been spared a labyrinthine
deviation, had the work of Berger come to acknowledgement
and fruition in his day. But again there was delay; Berger
waved the fairy wand in 1928; the transformation of Cinder-
ella was a process of years.
There were reasons for this delay. For one thing, Berger
was not a physiologist and his reports were vitiated by the
vagueness and variety of his claims and the desultory nature
of his technique. He was indeed a surprisingly unscientific
scientist, as personal acquaintance with him later confirmed.
The first occasion on which the possibilities of clinical elec-
troencephalography were discussed in England was quite an
informal one. It was in the old Central Pathological Labora-
tory at the Maudsley Hospital in London, in 1929. The team
there under Professor Golla was in some difficulty about
electrical apparatus; they were trying to get some records of
the "Berger rhythm," using amplifiers with an old galvanom-
eter that fused every time they switched on the current. Golla
was anxious to use the Matthews oscillograph, then the last
word in robust accuracy, to measure peripheral and central
conduction times. I was still working at Cambridge under the
watchful eye of Adrian and Matthews and was pleased to
introduce this novelty to him and at the same time, with
undergraduate superiority, put him right on a few other
points. When, at lunch around the laboratory table, he re-
ferred to the recent publication of Berger's claims, I readily
declared that anybody could record a wobbly line, it was a
string of artefacts, even if there were anything significant in
it there was nothing you could measure, and so on. Golla
agreed with milder scepticism, but added: "If this new ap-
paratus is as good as you say, it should be easy to find out
whether Berger's rhythm is only artefact; and if it isn't, the
frequency seems remarkably constant; surely one could meas-
ure that quite accurately." And he surmised that there would
be variations of the rhythm in disease.
Cambridge still could not accept the brain as a proper study
for the physiologist. The wobbly line did not convince us or
anybody else at that time. Berger's "elektrenkephalograms"
were almost completely disregarded. His entirely original and
painstaking work received little recognition until in May,
1934, Adrian and Matthews gave the first convincing demon-
stration of the "Berger rhythm" to an English audience, a
meeting of the Physiological Society at Cambridge.
Meanwhile, Golla was reorganising his laboratory, and his
confidence in the possibilities of the Berger method was
growing. When he invited me to join his research team as
physiologist at the Central Pathological Laboratory, my first
272
A Mirror for the Brain
task was to visit the German laboratories, including particu-
larly that of Hans Berger.
Berger, in 1935, was not regarded by his associates as in
the front rank of German psychiatrists, having rather the
reputation of being a crank. He seemed to me to be a modest
and dignified person, full of good humour, and as unperturbed
by lack of recognition as he was later by the fame it even-
tually brought him. But he had one fatal weakness: he was
completely ignorant of the technical and physical basis of
his method. He knew nothing about mechanics or electricity.
This handicap made it impossible for him to correct serious
shortcomings in his experiments. His method was a simple
adaptation of the electrocardiographic technique by which
the electrical impulses generated by the heart are recorded.
At first he inserted silver wires under the subject's scalp; later
he used silver foil bound to the head with a rubber bandage.
Nearly always he put one electrode over the forehead and
one over the back of the head; leads were taken from these to
an Edelmann galvanometer, a light and sensitive "string" type
of instrument, and records were taken by an assistant photog-
rapher. A potential change of one-ten-thousandth of a volt —
a very modest sensitivity by present standards — could just be
detected by this apparatus. Each record laboriously produced
was equivalent to that of two or three seconds of modern con-
tinuous pen recording. The line did show a wobble at about
10 cycles per second. ( See Figure 3. ) He had lately acquired
a tube amplifier to drive his galvanometer, and his pride and
pleasure in the sweeping excursions of line obtained by its
use were endearing.
Berger carried the matter as far as his technical handicap
permitted. He had observed that the larger and more regular
273
tr»
/eo.
§5
>--
*i
c ?
*i
•3 2
.C
ea .2
JLp
i3 t«
.»»fc
C to
o .
3°
w O
4-1
Av
u a
O l-<
*"
o-2
JL
-Q
I CO
Is
<«■
o <u
•4
r-s "0
.»<
O o
<£
^ S
4->
» J?
3 *
o
4 X
"8 s
Jf
"""J
01 «4H
*1
.StJ
2*"
03 <u
£ ^3 ^
O N^''0
"MO
1
T3T) O
0) 0) O
-<t»r
^ <-» Ih
-***!
<U £ 9>
.*l""v*
rC .a >
•VS.
HI- °
: m in
. « "55
-"r^
©3 ,w Ji*
I X
K"h id
4f*ra
tt <L> C
3 W)c9
—*.
bjD a; o
«■""
<
•
*o S
274
A Mirror for the Brain
rhythms tended to stop when the subject opened his eyes or
solved some problem in mental arithmetic. This was con-
firmed by Adrian and Matthews with leads from electrodes
on Adrian's head attached to a Matthews amplifier and ink-
writing oscillograph. This superior apparatus, and a more
careful location of electrodes, enabled them to go a step fur-
ther and prove that the 10 cycles per second rhythm arises in
the visual association areas in the occiput and not, as Berger
supposed, from the whole brain.
Only some years later was it realised what an important
step this was. Its significance could not be recognised while
so little was known about the components of the "wobbly
line," the electroencephalogram or, abbreviated, EEC Un-
avoidably at the time, the significance of the salient character
of the normal EEG was overlooked; it was found, in Adrian's
phrase, "disappointingly constant." The attention of many
early workers in electroencephalography therefore turned
from normal research to the study of nervous disease. In im-
mediate rewards this has always been a rich field. In this in-
stance, a surprising state was soon reached wherein what
might be called the electropathology of the brain was further
advanced than its electrophysiology.
In the pathological laboratory, Golla's earlier surmise, that
there would be variations of the rhythmic oscillation in
disease, was soon verified. A technique was developed there
by which the central point of the disturbance in the tissue
could be accurately determined. For surgery, the immediate
result of perfecting this technique was important; it made
possible the location of tumours, brain injuries, or other phys-
ical damage to the brain. It was helpful in many head cases
during the war as well as in daily surgical practice.
275
The study of epilepsy and mental disorders also began to
occupy the attention of many EEG workers. The difficulties
encountered in these subjects threw into prominent relief
the essential complexity of the problem as compared with
those of classical physiology. The hope of isolating single func-
tions had now been abandoned; those who entered this field
were committed to studying the brain as a whole organ and
through it the body as a whole organism. They were therefore
forced to multiply their sources of information.
It is now the general EEG practice, not only for clinical
purposes, but in research, to use a number of electrodes si-
multaneously, indeed as many as possible and convenient.
The standard make of EEG recorder has eight channels.
Eight pens are simultaneously tracing lines in which the
recordist, after long experience, can recognise the main com-
ponents of a complex graph. The graphs can also be auto-
matically analysed into their component frequencies. A more
satisfactory method of watching the electrical changes in all
the main areas, as in a moving picture, a much more informa-
tive convention than the drawing of lines, has been devised
at the Burden Neurological Institute. This will be described
after a simple explanation of what is meant by the rhythmic
composition of the normal EEG; for its nature, rather than the
methods of recording and analysing it, is of first importance
for understanding what follows.
If you move a pencil amply but regularly up and down on
a paper that is being drawn steadily from right to left, the
result will be a regular series of curves. If at the same time
the paper is moving up and down, another series of curves
will be added to the line drawn. If the table is shaking, the
vibration will be added to the line as a ripple. There will then
276
A Mirror for the Brain
be three components integrated in the one wavy line, which
will begin to look something like an EEG record. The line
gives a coded or conventional record of the various fre-
quencies and amplitudes of different physical movements. In
similar coded or integrated fashion the EEG line reports the
frequencies and amplitudes of the electrical changes in the
different parts of the brain tapped by the electrodes on the
scalp, their minute currents being relayed by an amplifier
to the oscillograph which activates the pens.
All EEG records contain many more components than this;
some may show as many as 20 or 30 at a time in significant
sizes. Actually there may be tens of thousands of impulses
woven together in such a manner that only the grosser com-
binations are discernible.
A compound curve is of course more easily put together
than taken apart. ( See Figure 4. ) The adequate analysis of a
few inches of EEG records would require the painstaking com-
putation of a mathematician — it might take him a week or so.
The modern automatic analyser in use in most laboratories
writes out the values of 24 components every 10 seconds, as
well as any averaging needed over longer periods.
The electrical changes which give rise to the alternating
currents of variable frequency and amplitude thus recorded
arise in the cells of the brain itself; there is no question of
any other power supply. The brain must be pictured as a
vast aggregation of electrical cells, numerous as the stars of
the Galaxy, some 10 thousand million of them, through which
surge the restless tides of our electrical being relatively thou-
sands of times more potent than the force of gravity. It is
when a million or so of these cells repeatedly fire together
277
that the rhythm of their discharge becomes measureable in
frequency and amplitude.
What makes these million cells act together — or indeed
what causes a single cell to discharge — is not known. We are
still a long way from any explanation of these basic mechanics
of the brain. Future research may well carry us, as it has car-
-2 -^-^^\^^v\AAAAAAAAAAA/VV\A/
■= v^aaAAAAAAAAAAAAAAA/
8 +9 +10
l"il"l,
cow TOmmmvfAmflmr/w^
8 ' [/J
9 V • " < »
n • ■
ii i ii
10
^Mmwrnmimmm
« ii mm t
I iniiml i . i I I i ■
*>*
*.*^-i>* *
— io sic.
Figure 4. "A compound curve is more easily put together than
taken apart." (a) A compound curve in which the three components
can be detected by visual inspection, ratios 1:2 and 2:3. (b) The three
components (ratios 8:9, 9:10) of this compound curve cannot be
determined at sight The bottom line shows their frequencies auto-
matically recorded every 10 seconds. Note the accidental similarity of
this curve to the EEG record of alpha rhythms in Figure 3 (b).
278
A Mirror for the Brain
ried the physicist in his attempt to understand the compo-
sition of our atomic being, into vistas of ever increasing
enchantment but describable only in the convention of mathe-
matical language. Today, as we travel from one fresh vista
to another, the propriety of the language we use, the con-
vention we adopt, becomes increasingly important. Arith-
metic is an adequate language for describing the height and
time of the tides, but if we want to predict their rise and fall
we have to use a different language, an algebra, with its spe-
cial notation and theorems. In similar fashion, the electrical
waves and tides in the brain can be described adequately by
counting, by arithmetic; but there are many unknown quan-
tities when we come to the more ambitious purposes of under-
standing and predicting brain behaviour — many x's and y's;
so it will have to have its algebra. The word is forbidding
to some people; but, after all, it means no more than "the
putting together of broken pieces."
EEG records may be considered, then, as the bits and
pieces of a mirror for the brain, itself speculum speculorum.
They must be carefully sorted before even trying to fit them
together with bits from other sources. Their information
comes as a conventional message, coded. You may crack the
code, but that does not imply that the information will neces-
sarily be of high significance. Supposing, for instance, you
pick up a coded message which you think may be about a mo-
mentous political secret. In the first stage of decoding it you
might ascertain that the order of frequency of the letters was
ETAONI. This does not sound very useful information; but
reference to the letter-frequency tables would assure you at
least that it was a message in English and possibly intelligible.
Likewise, we watch the frequencies as well as the amplitude
and origin of the brain rhythms, knowing that many earnest
seekers for the truth have spent lifetimes trying to decipher
279
what they thought were real messages, only to find that their
horoscopes and alembics contained gibberish. The scientist
is used to such hazards of research; it is only the ignorant
and superstitious who regard him, or think he regards him-
self, as a magician or priest who is right about everything all
the time.
Brain research has just about reached the stage where the
letter frequencies of the code indicate intelligibility and their
grouping significance. But there is this complication. The
ordinary coded message is a sequence in time; events in the
brain are not a single sequence in time — they occur in three-
dimensional space, in that one bit of space which is more
crowded with events than any other we can conceive. We
may tap a greater number of sectors of the brain and set more
pens scribbling; but the effect of this will only be to multiply
the number of code signals, to the increasing embarrassment
of the observer, unless the order and inter-relation of the
signals can be clarified and emphasised. Redundancy is al-
ready a serious problem of the laboratory.
The function of a nervous system is to receive, correlate,
store and generate many signals. A human brain is a mecha-
nism not only far more intricate than any other but one that
has a long individual history. To study such a problem in
terms of frequency and amplitude as a limited function of
time — in wavy lines — is at the best over-simplification. And
the redundancy is indeed enormous. Information at the rate
of about 3,600 amplitudes per minute may be coming through
each of the eight channels during the average recording pe-
riod of 20 minutes; so the total information in a routine record
may be represented by more than half a million numbers; yet
the usual description of a record consists only of a few sen-
tences. Only rarely does an observer use more than one-
hundredth of one per cent of the available information.
280
A Mirror for the Brain
"What's in a brain that ink may character . . . ?"
For combining greater clarity with greater economy, many
elaborations of methods have been adopted in clinic and
laboratory. They still do not overcome the fundamental em-
barrassment of redundancy and the error of over-simphfi ca-
tion, both due to the limitations of a time scale. A promising
alternative is a machine that draws a snapshot map instead
of a long history, projecting the electrical data visually on a
spatial co-ordinate system which can be laid out so as to repre-
sent a simple map or model of the head. This moving pano-
rama of the brain rhythms does approximate to Sherrington's
"enchanted loom where millions of flashing shuttles weave a
dissolving pattern, always a meaningful pattern though never
an abiding one." ( Figure 5. )
We have called the apparatus which achieves this sort of
effect at the Burden Institute a toposcope, by reason of its
display of topographic detail. The equipment was developed
by Harold Shipton, whose imaginative engineering trans-
formed the early models from entertainment to education.
Two of its 24 channels are for monitoring the stimuli; the
others, instead of being connected with pens, lead the elec-
trical activity of the brain tapped by the electrodes for display
on the screens of small cathode-ray tubes. So instead of wavy
lines on a moving paper, the observer sees, to quote Sherring-
ton again, "a sparkling field of rhythmic flashing points with
trains of travelling sparks hurrying hither and thither." As-
sembled in the display console, 22 of the tubes give a kind
of Mercator's projection of the brain. Frequency, phase and
time relations of the rhythms are shown in what at first ap-
pears to be a completely bewildering variety of patterns in
each tube and in their ensemble. Then, as the practised eye
gains familiarity with the scene, many details of brain ac-
tivity are seen for the first time. A conventional pen machine
281
is simultaneously at the disposal of the observer, synchronised
so that, by turning a switch, a written record of the activity
seen in any five of the tubes can be made. Another attachment
is a camera with which at the same time permanent snapshot
records of the display can be obtained. ( Figure 6. )
Thus, from Berger's crude galvanometer to this elaborate
apparatus requiring a whole room of its own, electroenceph-
alography has progressed from a technique to a science. Its
clinical benefits, by-products of free research, are acknowl-
edged; they can be gauged by the vast multiplication of EEG
laboratories. From Berger's lone clinic have sprung several
hundred EEG centres — more than 50 in England alone. Lit-
erally millions of yards of paper have been covered with fran-
tic scribblings. In every civilized country there is a special
learned society devoted to the discussion of the records and
to disputation on technique and theory. These societies are
banded together in an International Federation, which pub-
lishes a quarterly Journal and organises international con-
gresses.
For a science born, as it were, bastard and neglected in
infancy, this is a long way to have travelled in its first quarter
of a century. If it is to provide the mirror which the brain
requires to see itself steadily and whole, there is still a long
road ahead. The following chapters give the prospect as seen
from the present milestone, assuming that such studies are
allowed to continue. Looking back, we realise that the present
scale of work as compared with previous physiological re-
search is elaborate and expensive. But our annual cost of con-
ducting planned investigations of a fundamental nature into
man's supreme faculties is less than half that of one medium
tank, and the money spent on brain research in all England
is barely one-tenth of one per cent of the cost of the national
mental health services.
282
A Mirror for the Brain
Figure 5. ". . . a moving panorama of the brain rhvthms." The Toposcope Laboratory. The subject's couch and
triggered stroboscope (flicker) reflector at extreme left be\ond desk of 6-channel pen recorder with remote control
panel. The 22-channeI toposcope amplifier is in the background, the display panel at right centre, camera and pro-
jector at extreme right.
Figure 6. ". . . always a meaningful pattern though never an abiding one."
Snapshots of the "sparkling field of rhythmic flashing points." Each of the tube
screens, which form a chart oi the head seen from above with nose at top.
shows bv the flashing sectors of its disc the activity of the corresponding area
of the brain. (Top left) Resting alpha rhythms, (fop right) Theta rhythms in
anger. (Bottom left) Wide response to double flashes of light. (Bottom right)
Spread of response to triple flashes.
283
Physics is full of concepts of which we cannot form
simple pictures. Therefore the authors, like most
modern scientists, recommend taking a "mathematical
view. "
19 Scientific Imagination
Richard P. Feynman, Robert B. Leighton, and Matthew Sands
Excerpt trom The Feynman Lectures on Physics, Volume II, 1964.
I have asked you to imagine these electric and magnetic fields. What do you
do? Do you know how? How do /imagine the electric and magnetic field? What
do / actually see? What are the demands of scientific imagination? Is it any
different from trying to imagine that the room is full of invisible angels? No, it is
not like imagining invisible angels. It requires a much higher degree of imagination
to understand the electromagnetic field than to understand invisible angels. Why?
Because to make invisible angels understandable, all I have to do is to alter their
properties a little bit— I make them slightly visible, and then I can see the shapes
of their wings, and bodies, and halos. Once I succeed in imagining a visible angel,
the abstraction required — which is to take almost invisible angels and imagine
them completely invisible— is relatively easy. So you say, "Professor, please give
me an approximate description of the electromagnetic waves, even though it may
be slightly inaccurate, so that I too can see them as well as I can see almost invisible
angels. Then I will modify the picture to the necessary abstraction."
I'm sorry I can't do that for you. I don't know how. I have no picture of this
electromagnetic field that is in any sense accurate. I have known about the electro-
magnetic field a long time— I was in the same position 25 years ago that you are
now, and I have had 25 years more of experience thinking about these wiggling
waves. When I start describing the magnetic field moving through space, I speak
of the E- and # fields and wave my arms and you may imagine that I can see them.
285
I'll tell you what I see. I see some kind of vague shadowy, wiggling lines — here
and there is an E and B written on them somehow, and perhaps some of the lines
have arrows on them — an arrow here or there which disappears when I look too
closely at it. When I talk about the fields swishing through space, I have a terrible
confusion between the symbols I use to describe the objects and the objects them-
selves. I cannot really make a picture that is even nearly like the true waves. So
if you have some difficulty in making such a picture, you should not be worried
that your difficulty is unusual.
Our science makes terrific demands on the imagination. The degree of
imagination that is required is much more extreme than that required for some of
the ancient ideas. The modern ideas are much harder to imagine. We use a lot
of tools, though. We use mathematical equations and rules, and make a lot of
pictures. What I realize now is that when I talk about the electromagnetic field in
space, I see some kind of a superposition of all of the diagrams which I've ever
seen drawn about them. I don't see little bundles of field lines running about be-
cause it worries me that if I ran at a different speed the bundles would disappear.
I don't even always see the electric and magnetic fields because sometimes I think
I should have made a picture with the vector potential and the scalar potential,
for those were perhaps the more physically significant things that were wiggling.
Perhaps the only hope, you say, is to take a mathematical view. Now what is
a mathematical view? From a mathematical view, there is an electric field vector
and a magnetic field vector at every point in space; that is, there are six numbers
associated with every point. Can you imagine six numbers associated with each
point in space? That's too hard. Can you imagine even one number associated
with every point? I cannot! I can imagine such a thing as the temperature at every
point in space. That seems to be understandable. There is a hotness and coldness
that varies from place to place. But I honestly do not understand the idea of a
number at every point.
So perhaps we should put the question: Can we represent the electric field by
something more like a temperature, say like the displacement of a piece of jello?
Suppose that we were to begin by imagining that the world was filled with thin
jello and that the fields represented some distortion — say a stretching or twisting —
of the jello. Then we could visualize the field. After we "see" what it is like we
could abstract the jello away. For many years that's what people tried to do.
Maxwell, Ampere, Faraday, and others tried to understand electromagnetism
this way. (Sometimes they called the abstract jello "ether.") But it turned out that
the attempt to imagine the electromagnetic field in that way was really standing in
the way of progress. We are unfortunately limited to abstractions, to using in-
struments to detect the field, to using mathematical symbols to describe the field,
etc. But nevertheless, in some sense the fields are real, because after we are all
finished fiddling around with mathematical equations— with or without making
pictures and drawings or trying to visualize the thing— we can still make the instru-
ments detect the signals from Mariner II and find out about galaxies a billion miles
away, and so on.
286
Scientific Imagination
The whole question of imagination in science is often misunderstood by people
in other disciplines. They try to test our imagination in the following way. They
say, "Here is a picture of some people in a situation. What do you imagine will
happen next?" When we say, "I can't imagine," they may think we have a weak
imagination. They overlook the fact that whatever we are allowed to imagine in
science must be consistent with everything else we know: that the electric fields and
the waves we talk about are not just some happy thoughts which we are free to
make as we wish, but ideas which must be consistent with all the laws of physics
we know. We can't allow ourselves to seriously imagine things which are obviously
in contradiction to the known laws of nature. And so our kind of imagination is
quite a difficult game. One has to have the imagination to think of something that
has never been seen before, never been heard of before. At the same time the
thoughts are restricted in a strait jacket, so to speak, limited by the conditions that
come from our knowledge of the way nature really is. The problem of creating
something which is new, but which is consistent with everything which has been
seen before, is one of extreme difficulty.
While I'm on this subject I want to talk about whether it will ever be possible
to imagine beauty that we can't sec. It is an interesting question. When we look
at a rainbow, it looks beautiful to us. Everybody says, "Ooh, a rainbow." (You
see how scientific I am. I am afraid to say something is beautiful unless I have an
experimental way of defining it.) But how would we describe a rainbow if we were
blind? We are blind when we measure the infrared reflection coefficient of sodium
chloride, or when we talk about the frequency of the waves that are coming from
some galaxy that we can't see — we make a diagram, we make a plot. For instance,
for the rainbow, such a plot would be the intensity of radiation vs. wavelength
measured with a spectrophotometer for each direction in the sky. Generally, such
measurements would give a curve that was rather flat. Then some day, someone
would discover that for certain conditions of the weather, and at certain angles in
the sky, the spectrum of intensity as a function of wavelength would behave
strangely; it would have a bump. As the angle of the instrument was varied only a
little bit, the maximum of the bump would move from one wavelength to another.
Then one day the physical review of the blind men might publish a technical article
with the title "The Intensity of Radiation as a Function of Angle under Certain
Conditions of the Weather." In this article there might appear a graph such as
the one in Fig. 20-5. The author would perhaps remark that at the larger angles
there was more radiation at long wavelengths, whereas for the smaller angles the
maximum in the radiation came at shorter wavelengths. (From our point of view,
we would say that the light at 40° is predominantly green and the light at 42° is
predominantly red.)
287
Wavelength
Fig. 20-5. The intensity of electro-
magnetic waves as a function of wave-
length for three angles (measured from
the direction opposite the sun), observed
only with certain meteorological con-
ditions.
Now do we find the graph of Fig. 20-5 beautiful? It contains much more de-
tail than we apprehend when we look at a rainbow, because our eyes cannot see
the exact details in the shape of a spectrum. The eye, however, finds the rainbow
beautiful. Do we have enough imagination to see in the spectral curves the same
beauty we see when we look directly at the rainbow? I don't know.
But suppose I have a graph of the reflection coefficient of a sodium chloride
crystal as a function of wavelength in the infrared, and also as a function of angle.
I would have a representation of how it would look to my eyes if they could see
in the infrared — perhaps some glowing, shiny "green," mixed with reflections from
the surface in a "metallic red." That would be a beautiful thing, but I don't know
whether I can ever look at a graph of the reflection coefficient of NaCl measured
with some instrument and say that it has the same beauty.
On the other hand, even if we cannot see beauty in particular measured results,
we can already claim to see a certain beauty in the equations which describe general
physical laws. For example, in the wave equation (20.9), there's something nice
about the regularity of the appearance of the x, the y, the z, and the /. And this
nice symmetry in appearance of the x, y, z, and / suggests to the mind still a greater
beauty which has to do with the four dimensions, the possibility that space has
four-dimensional symmetry, the possibility of analyzing that and the developments
of the special theory of relativity. So there is plenty of intellectual beauty asso-
ciated with the equations.
a2i|>
3x2
ay
+ d2^ _ JL 924>
3z2 c2 3t2
(20.9)
288
Magnifying glasses, spectacles, cameras, projectors, eyes,
microscopes, telescopes— they all work on the same simple
principles.
20 Lenses and Optical Instruments
Physical Science Study Committee
From the textbook Physics by Physical Science Study Committee, 1 962.
Optical instruments — cameras, projectors, tele-
scopes, and microscopes — usually are built with
lenses ; that is, with pieces of refractive materials
to converge or diverge light according to our
design. A whole industry is devoted to the design
and production of such instruments or their
components. All these instruments are under-
stood and designed in terms of Snell's law. The
whole field of applications rests on the simple sum-
mary of refraction that we reached in the last
chapter, nx sin 6X = n2 sin d2. Most optical tech-
nology stems from this little bit of physics.
In this chapter, we want to learn how the laws
of refraction are related to the construction of
lenses and optical systems. An extensive treat-
ment of the design of optical systems is, however,
beyond the purpose of this chapter.
14-1. The Convergence of Light by a Set
of Prisms
We found in Chapter 12 that we could control
and redirect light beams by the use of curved
mirrors. Devices that can accomplish similar
purposes through refraction, instead of reflection,
are called lenses. To understand how a lens
operates, let us examine the behavior of light in
passing through the combination of a plate of
glass with parallel sides and the two triangular
prisms shown in Fig. 14-1 (a). If a parallel beam
of light falls on this system from the left, so that
it is normally incident on the plate of glass, it will
behave as indicated by the rays shown in the
figure. The light that passes through the plate
in the center will continue along its original
direction, since the angle of incidence is 0°.
Light striking the upper prism will be deviated
downward by an amount depending on the open-
ing angle of the prism and on its index of refrac-
14-1. Construction of a lens by the process of subdividing prismatic
sections.
289
tion. Similarly, light striking the lower prism
will be deviated upward. As a result, there is a
region, shown shaded in the figure, through
which passes almost all of the light that falls on
the plate and the prisms.
The convergence of a parallel beam of light into
a limited region by this system resembles the
convergence of a similar beam by a set of mirrors.
(See Section 12-6.) While working with mirrors,
we decreased the size of the region into which
the light was converged by using an increased
number of mirrors, each smaller than the original
one. Let us try the same scheme here. Fig.
14-1 (b) shows parts of the central plate and of
the two prisms cut away and replaced by pieces
of new prisms. The size of the shaded region is
clearly smaller than it was before.
If we continue the process of removing parts
of the prisms and replacing them by sections
having smaller opening angles, we come closer
and closer to a piece of glass with the smoothly
curved surface shown in Fig. 14-1 (c). This
device is the limit that is approached as we
increase the number of prisms indefinitely, just as
the parabolic mirror of Fig. 12-16 was the limit
approached as we used more and more plane
mirrors to converge parallel light. In Fig. 14-2
we have actually carried out the construction
indicated in Fig. 14-1. The lens produced by
the process that we have outlined converges all
of the parallel light that strikes it to a line as
shown in Fig. 14-3.
14-2. The experiments diagramed in Fig. 14—1.
14—3. Convergence of light by a cylindrical lens like the one
shown in Fig. 14—1. Note that the light is brought to a focus
along a line.
14-2. Lenses
The device we have just constructed is called a
cylindrical lens. Notice that we have not given
any definition of the surfaces of the lens, except
that they are obtained by increasing indefinitely
the number of sections of prisms that are used to
converge the light. It is possible to show that
these surfaces are approximated very closely by
circular cylinders. In other words, the lines
representing the surfaces in Fig. 14-1 (c) are arcs
of circles. The differences between the ideal
surfaces and those of circular cylinders are very
slight whenever both the width of the lens and its
maximum thickness are considerably smaller
than the distance from the lens to the line at which
parallel light is converged.
Cylindrical lenses bring the light from a distant
point source of light to a focus along a line. For
most purposes we prefer that the light from a point
source should be focused at a point. This focus-
ing can be accomplished by constructing a lens
whose surfaces curve equally in all directions.
Such surfaces are portions of spheres. Almost
all lenses are bounded by two spherical surfaces.
The line passing through the center of the lens
and on which the centers of the two spheres are
located is called the axis of the lens. The point
on this axis at which incident parallel rays focus
or converge is the principal focus, F. The distance
of the principal focus from the center of the lens
is known as the focal length, f
The two surfaces of a lens do not always have
the same radius. For example, the lens shown in
Fig. 14-4 has a spherical surface of much larger
290
Lenses and Optical Instruments
^4&Z?
L -~^
■^^"^^— "
*- — ~^^^^Zs^
t r
\-
14—4. A lens with surfaces of unequal radii.
radius at its right-hand boundary than it has at
the left.
If a lens is thin compared to its focal length,
it makes no difference which side of the lens the
light enters, the focal length is always the same.
This symmetry is obvious if the lens is itself
symmetric. That it is true for all thin lenses can
easily be shown by an experiment in which a lens
is used to focus the parallel rays of the sun to a
point on a piece of paper or cardboard. If the
lens is then flipped over, the focus occurs at the
same distance from the lens (Fig. 14-5).
This result is also predicted by a detailed
application of Snell's law from which we find
H»-<-4>
/
where R^ and R2 are the radii of the opposing
spherical surfaces.* We see that interchanging
Ri and R2, which is equivalent to turning the lens
aver, does not change the calculated value of/.
From this equation, we can also see that when
Ri and R2 are small, the lens will have a short
focal length. This is illustrated in Fig. 14-6 where
ive see that the paths of light rays through the
[ens in (b) are bent more sharply, so that the
focal length is shorter than in (a).
14-3. Real Images Formed by Lenses
We have thus far concentrated our attention
>n the focusing of light by a lens when the light
x>mes from a very distant object. In the practical
lse of lenses, we are commonly interested in the
'We shall not give the proof of this "lens maker's" formula
here. Although no new physics is involved, the proof is a
long-winded application of trigonometry and Snell's law.
Later, however, we can use the results of further study to
get the formula more easily. It is therefore discussed at
the end of Part II (see pages 302-303).
14—5. The principal foci of a lens. For thin lenses the focal
distance is the same for parallel light entering either the side with
a small radius (a) or the side with a larger radius (b).
light coming from near-by objects and we all
know that lenses do form images of such objects.
We can locate the images with the help of the
knowledge that we have gained about the be-
havior of initially parallel rays.
Fig. 14-7 shows a lens, an object H», and its
image Ht. To find the location of this image, we
draw the two principal rays from the top of the
object, one ray parallel to the axis and the other
through the principal focus F2. The ray parallel
to the axis is bent by the lens so as to pass through
the principal focus Fv We also know that rays
coming from the right and parallel to the axis
would be deviated to pass through the other
principal focus F2. It follows from the reversi-
bility of light paths that the ray from the top of
H0 that passes through F2 from the left must travel
parallel to the axis after it has passed through the
lens. All rays starting from the top of H0 will
converge very close to the point at which these
two bent rays intersect. This point is therefore
the real image of the top of H0.
We could have chosen any other point on the
object and located its image in the same way.
Had we done so for a number of points, we would
have found that the image, //;, falls along the
line that is shown in the figure.
291
14—6. The shorter the radius of the surface of a lens, the shorter
the focal length.
You probably have noticed that, in constructing
the two principal rays, we have not considered the
exact path of the ray within the lens, but have
broken it sharply. This approximate construction
is good enough for our present purposes because
our location of the two principal foci is accurate
only if the lens thickness (at its center) is small
compared with the focal length. The only lenses
to which our construction accurately applies are
therefore thin lenses. For the purposes of ray
diagrams, we may consider such lenses to be
circular plates perpendicular to the axis.
Convex lenses, like parabolic mirrors, focus
parallel rays to a point. Lenses, therefore, obey
the same equation relating image distance, focal
length, and object distance as do mirrors :
The proof of this equation in the case of convex
lenses is the same as for mirrors (Section 12-9).
As there, we use the shaded similar triangles
formed by the principal rays shown in Fig. 14-7.
Considering first the shaded similar triangles to
the left of the lens, we see that H-JH0 =//50-
The shaded triangles to the right of the lens give
HJH0 = SJf- Combining the two equations, we
have
S0S{ =p.
14-4. Camera, Projector, and Eye
Produce an image of the sun with a convex lens.
Since the sun is far away, the image is formed
practically at the principal focus and you can see
it there on a piece of paper. Images of closer
objects lie beyond the principal focus; and, in
order to capture them on paper or on a photo-
graphic film, we have to change the distance
between lens and film. To make a photographic
camera, then, we usually make a light-tight box
with a bellows that allows us to move the lens.
By adjusting the length of the bellows, we can
place a sharp image on the photographic film.
With some cameras we can place a piece of
ground glass where the film is later inserted.
14-7. The formation of a real image by a converging lens.
292
Lenses and Optical Instruments
14-8. A camera. The ray* of light that form the image of the
head of the arrow are indicated.
We can then view the image directly and focus
sharply on the particular object we want to
photograph (Fig. 14-8).
As long as the object is more than twice the
focal distance from the lens, so that S0 is longer
than/, the image size is smaller than the object,
as Hi/H0 =f/S0 shows. When a small object is
brought closer to the principal focus, the image
moves to distances behind the lens that are large
retina
optic nerve
retina
optic nerve
14—9. The lens of an eye adjusted to focus the light from a dis-
tant object (a) and from one near by.
compared with the focal length; also, the image
becomes bigger than the object. Consequently,
to photograph small objects, a lens of short focal
length is useful.
A projector is just a camera worked backwards.
You can make one by taking the back off a
camera, mounting the slides or film where the
film usually goes, and shining a bright light
through the film and out through the lens. The
lens then forms an enlarged image well in front
of the camera, where you can place a screen.
In cameras, projectors, and other man-made
optical instruments, images are always brought
into focus by changing the position of a lens with
respect to the object. The eye, on the other hand,
is unusual: it focuses images on the retina by
changing its curvature and hence the focal length
of its lens. When an object is at a very large
distance from the eye, the rays entering the eye
are nearly parallel and an image is formed at the
principal focus as shown in Fig. 14-9 (a). When
a close-by object is viewed, the image is formed
beyond the focal point, and eye muscles form
the elastic eye lens into a sharper curve, de-
creasing its focal length so that a real image
will form on the retina [Fig. 14-9 (b)].
14-5. The Magnifier or Simple Microscope
Let us go back to the small object that we
brought close to the principal focus of a lens.
As the object is moved through the principal
focus the real image moves infinitely far away
on the other side of the lens ; and when the object
is between the lens and the principal focus a virtual
image is formed behind the object just as in the
case of a concave mirror that we discussed in
293
Section 12-10. The situation is illustrated in
Fig. 14-10. As in the case of the concave mirror,
the convex lens always forms an enlarged virtual
image.
What is the maximum magnification that we
can obtain in this way? If we wish to see the
greatest possible detail in an object, we get it as
close to the eye as possible, thus giving a large
real image on the retina of the eye. But there is
a limit to how close we can view an object. As
the object gets closer to the eye, the eye muscles
must change the shape of the eye lens so that its
radius of curvature becomes smaller and smaller
in order to form a sharply focused real image on
the retina. Soon a limit is reached; the adult
eye cannot accommodate to an object closer than
about 25 cm. This object distance is called the
distance of most distinct vision. Try bringing
a pencil closer and closer to your eye. You will
see more and more detail until finally, with a great
straining of your eye muscles, you can no longer
keep a sharp image. Is your distance of most
distinct vision greater or less than the average
of 25 cm?
A convex lens helps us to see more detail by
forming an enlarged virtual image which we can
place at a comfortable distance from the eye.
We notice in Fig. 14-10 that no matter where
the object is placed between the lens and F2, the
top of the image always lies on the line FYD, and
Hi = — Si as usual.
Consequently, to make the image look as large
as possible we should bring our eye right up to
the lens as in Fig. 14-11; and in addition we
should move the object (or the lens and our eye)
until the image gets as close as we can clearly
accommodate. This is the way to get the largest
angle between the rays entering our eye from the
top and from the bottom of the object; and since
this light is what the eye works with, it is the
way to make the object (or its virtual image)
look largest.
Now for our own comfort we place the image
at the distance of most distinct vision d, so the
image distance S, (measured from Fx) is approxi-
mately given by Si = d +f. Therefore
Hi -5?S, - y°(</+/) = H0(j+ l)-
Furthermore, since we are looking at this image
14-10. Formation of a virtual image by a converging lens .
from the distance of most distinct vision just as
we could best look at the object without the aid
of the lens, the magnification of the image we
see is Hi/H0. That is, maximum magnification is
This equation tells us the greatest magnification
of a simple microscope. What, then, determines
how great a magnification we can get? The focal
length, /, of the convex lens is the determining
factor; the smaller it is, the greater the magnifica-
tion. In order to get a small/we use glass of high
refractive index to produce sharper bending of
the light for a given curvature of the lens surfaces.
Also we need surfaces of small radius (sharp
curvature). But a small radius of curvature means
a small lens size, since the lens diameter cannot
1
distance of
most distinct vision
14-11. A converging lens used as a magnifier. The image is
placed at the distance of closest distinct vision. Since the eye
is very close to the lens, the distance from the image to the lens
is about the same as that to the eye.
294
-•::•::*•: in: Bjpfied kNflMNHt
onijrainrt nirrnxrsipt. Tm *»v»pitMi inrt inntrchv*
cut tr hit mi:
"* ■-■,:'.;■■-.:
Tie hck :r.m:~ . . : -
-,-
': :u :;i
■-
:•.
:r
:.:: :;:ot
nu
■;::
14—13. Distortion by a lens. These three photographs were
made by looking through the same lens. At left, the lens was
held so that the page of the telephone book is slightly below
the focal region; in the middle picture, the page is in the focal
region; at right, the focal region lies below the page. Note
the geometrical distortions.
14-7. Limitations of Optical Instruments
If you hold a magnifying glass close to this
page, you can see a clear, undistorted and slightly
magnified image of the print. Now slowly raise
the glass from the paper and at the same time
increase the distance between your eyes and the
lens. At some positions the image appears
distorted. If your eyes are far enough from the
glass, you may also detect some rainbow colors
when looking at a corner of the page. In Figure
14-13 we see three pictures. They were made
by looking through the same lens. In the first
picture the lens was held so that the page of the
telephone book lies entirely (but slightly) below
the focal region; in the second picture the page
is in the focal region; and in the last the focal
region lies below the page. Clearly, in each
case the image of the page looks quite different
from the page itself. Part of the game of design-
ing really good optical instruments is to minimize
the geometrical distortions so apparent in these
pictures.
What are the origins of these defects in images?
First, we know even for mirrors (Chapter 12)
that a surface designed to bring light from one
small object to a sharp focus is not the correct
surface to bring light to an exact focus from an
object at a different place. The same is true for
lens surfaces. Some blurring of the image there-
fore results. In addition, when we look through
different parts of a lens the images are at different
positions (and the magnification is different).
The image therefore is distorted. In photography
distortion and blurring are often cut down by
using a "stop," a barrier with a small hole in it
so that we use only a selected portion of the lens.
The colored edges of images usually arise
because of the dispersion of the light that passes
through a lens. The focal length of a lens is
slightly longer for red light than it is for blue
light, because the blue light is refracted more
strongly than the red. This undesirable effect is
called chromatic aberration. It can be greatly
reduced by using a weakly diverging lens made of
glass for which the index of refraction changes
greatly with color, in conjunction with a strongly
14-14. A lens built of two pieces to minimize the different
focal properties of different colors. Such doublets are often made
with one common surface and glued together. They are called
achromatic lenses.
196
Lenses and Optical Instruments
converging lens of glass for which the index of
refraction changes less with color (Fig. 14-14).
This trick makes the focal properties of the whole
system of lenses nearly the same for all colors.
The problem of designing a system of lenses
with the smallest amount of distortions and
aberrations is a very complicated one. But the
complications arise only in the detailed applica-
tions of the laws of refraction; they involve no
new principle. Disentangling these complica-
tions will not enrich our understanding of basic
optical phenomena, and therefore we shall not
do it here.
There is one limiting factor affecting optical
magnifiers which causes a blur in the image and
is of a fundamental nature. This is the inevitable
diffraction which results from the limited size
of the objective lens through which the light
must pass. At high magnification it is this
blurring that prevents us from seeing finer and
finer detail. We shall learn more about diffrac-
tion in Chapter 19.
FOR HOME, DESK, AND LAB
14-15. For Problem 1.
1. A crude converging lens can be constructed by
placing two 30°-60°-90° glass prisms together
with a glass block as shown in Fig. 14-15.
(a) What is the focal length of this "lens" to
one significant figure?
(b) Would such a lens form a clear image?
Explain.
2. If two 45° prisms of glass (index = 1.50) are ar-
ranged as in Fig. 14—16 they will not converge
parallel light. Why not? What will happen to
the light?
3. Some lighthouses and light buoys mark the posi-
tions of dangerous rocks and shoals. The light
must be concentrated at a low angle with respect
to the horizon (light directed upward is wasted)
and must be equally visible from all points of the
compass.
(a) Can you design a "lens" which will do this?
(b) Instead of using a continuous curved sur-
face, such lights often use a lens made of sections
of prisms. Can you draw a diagram of such a
lens? It is called a Fresnel lens after the French
physicist who first devised such a lens.
(c) Automobile headlights are constructed to
give a wide, flat, horizontal beam. Parabolic
reflectors are made to give a narrow beam which
passes through a Fresnel lens in the front of the
headlight. Examine an automobile headlight and
see if you can understand how it gives broad,
horizontal beams.
14-16. For Problem 2.
4. Use the Lens Maker's Formula
}--»ft+fi
to find the focal length of a glass iens (n = 1.50)
with one flat surface and one with a radius of 10
cm. (Such a lens is called a plano-convex lens.)
(a) What are the focal lengths of the two lenses
shown in Fig. 14-17? (Index of glass = 1.50.)
(b) How does the focal length of (b) compare
with a flat block of glass?
A lens (index = 1.50) has a focal length in air of
20.0 cm.
(a) Is its focal length in water greater or less
than in air?
(b) What is its focal length in water?
Hint: Notice that every individual refraction
depends on the relative index of refraction.
297
7. A lens whose focal length is 10 cm is used in a
slide projector to give a real image on a screen at a
distance of 6.0 meters.
(a) What will be the magnification?
(b) How far is the lens placed from the slide?
8. Prove that if two identical converging lenses of
focal length 10 cm are placed 40 cm apart, the
combination will form an upright image of an
object that is 20 cm away from the first lens and
the magnification will be 1.
9. (a) Prove that the size of the image of the sun
produced by a convex lens is proportional to the
focal length. What is the constant of pro-
portionality?
(b) How large an image of the sun (diameter
1.4 X 109 m) will be formed by a lens of focal
length 1.0 meter?
(c) What will be the ratio of the size of the
images of the sun formed by a lens of 10 cm focal
length and a lens of 10 m focal length?
10. How large an image will be formed of an artifi-
cial satellite (53 cm in diameter) passing at an
altitude of 500 miles, if it is photographed with a
camera whose focal length is 10 cm? Would you
expect an actual photograph to show a larger or
smaller image than the size you have calculated?
14-17. For Problem 5.
11. (a) What is the focal length of the lens in Fig.
14-18? (Index of the glass is 1.50.)
(b) By sketching the paths of some light rays,
show what the lens does to incident light parallel
to its axis.
(c) From the ray diagram of Fig. 14-18 (b),
show that S0Si =/2. Notice from which focal
points S0 and S; are measured.
(d) What happens as you move the object to-
ward the lens? Can S; ever get bigger than S0?
Is the image ever bigger than the object?
(e) How would you find (experimentally) the
focal length of a diverging lens?
12. Assume your distance of most distinct vision is
1 5 cm. What is the maximum magnification that
can be obtained with each of the following convex
lenses when used as a magnifying glass or simple
microscope?
(a) / = 30 cm,
(b) / = 10 cm,
(c) / = 1 cm,
(d) / = 1 mm.
(e) Graph the
a function of "/."
maximum magnification as
13. Assume your distance of most distinct vision is
25 cm. A compound microscope has an eyepiece
of 2.0 cm focal length and an objective of 4.0 mm
focal length. The distance between objective and
eyepiece is 22.3 cm. What is its magnification to
two significant figures?
14. Using the microscope of Problem 13, with the
same adjustment, we see an amoeba. With a
ruler, we measure the size of the virtual image by
looking at it with one eye and at the ruler with
the other. On the ruler the amoeba appears to be
about 10 cm long. About how big is it really?
15. For the maximum magnification of an eyepiece, we
d
found - + 1 where d is taken as the distance of
most distinct (or closest distinct) vision. If your
eyes can accommodate to see distinctly at 15 cm,
15 cm , ._
we should write — - — + 1 as the magnification
of a simple magnifier for you. Also, if you can
14-18. For Problem 11.
298
Lenses and Optical Instruments
object
<i
5 cm
10 cm
10 cm
14-19. For Problem 19.
accommodate to images no closer than 35 cm,
— j h 1 would apply. Why does the magnifi-
cation go up for someone who accommodates
poorly at small distances? Does he see more
detail than someone who can accommodate
closer? Be prepared to discuss this question in
class.
16. Two lenses both have a focal length of 20 cm, but
one has a diameter four times that of the other.
Draw sketches of the two lenses and tell how the
images they form differ.
17. (a) What is the ratio of the focal lengths of a
crown-glass lens for violet light and for red light?
(The index of refraction for various colors is
given in Table 4, Chapter 13.)
(b) Is the ratio the same for all kinds of glass?
18. A lens of focal length 20 cm is placed 30 cm from
a plane mirror and an object is placed on the axis
10 cm from the mirror. Where will the image of
the object be found?
19. Where are the images of the object in Fig. 14-19?
Can you see all the images if you look through the
lens
(a) with your eye near the lens?
(b) with your eye far from the lens?
FURTHER READING
Rogers, Frances, Lens Magic. Lippincott, 1957.
A history of the development of lenses, and a de-
scription of their many applications.
Texereau, Jean, How to Make a Telescope. Inter-
science Publishing Co., 1957.
Thompson, Allyn J., Making Your Own Telescope.
Sky Publishing Co., Cambridge, Mass., 1947.
Wald, George, "Eye and Camera." Scientific
American, August, 1950 (p. 32).
299
21
Everybody, however intelligent, has a mental block about
some aspect of life. This article from a British magazine of
humor, describes how electricity ought to behave.
Baffled!
Keith Waterhouse
Article in Punch, 1968.
YOU learn something new every day. With no thought of self-
improvement, for example, I was reading that story of Thurber's
in which he recalls his mother's belief that electricity leaks out
of an empty light socket if the switch has been left on. From
this I gathered — going by the general context, and the known
fact that Thurber was a humorist — that it doesn't.
I picked up another piece of electrical knowledge in 1951,
while working as a drama critic on the Yorkshire Evening Post.
Wanting to imply that a certain actress had given a muted
performance, I wrote that while undoubtedly she had an electric
presence, on this occasion it was as if the electricity had been
immersed in water. A kindly sub-editor explained to me that
when electricity gets wet, by some miracle of the elements it
intensifies rather than diminishes. I have never seen the sense
of this, but I conceded the point and have used only gas-driven
metaphor since that date.
I was never taught electricity at school, nor was it often a
topic of dinner-table conversation among my parents. What I
know about the subject I have mastered the hard way. Take, as
an instance, television, an electrical device of awesome com-
plexity. Unlike more privileged students, who are able to go
running to m'tutor every time the framehold goes wobbly, I
have had to learn in the School of Life that on the large rented
model the knobs are on the front whereas on the HMV portable
they are on the side. Similarly with electric irons. When I bought
my first electric iron there was no plug attached, presumably in
case I wanted to wind the flex around my neck and jump off
Westminster Bridge with it. There was a leaflet explaining how
to get the plug on, but this was of course in German, the inter-
national language of the household appliances industry. Only by
putting my natural intelligence to the problem did I eventually
work out the solution— find a German-speaking electrician.
And so, what with having perforce to change a light bulb
301
here and tune in a transistor radio there, I have picked up a
pretty sound working knowledge of electrical matters. It is not
comprehensive, God knows — I still can't fully understand why
you can't boil an egg on an electric guitar — but when I jot down
a summary of what I have learned, I marvel that I have never
been asked to write for the Electrical Journal:
1. Most electricity is manufactured in power stations where
it is fed into wires which are then wound around large drums.
2. Some electricity, however, does not need to go along wires.
That used in portable radios, for example, and that used in
lightning. This kind of electricity is not generated but is just
lying about in the air, loose.
3. Electricity becomes intensified when wet. Electric kettles
are immune to this.
4. Electricity has to be earthed. That is to say, it has to be
connected with the ground before it can function, except in the
case of aeroplanes, which have separate arrangements.
5. Electricity makes a low humming noise. This noise may be
pitched at different levels for use in doorbells, telephones,
electric organs, etc.
6. Although electricity does not leak out of an empty light
socket, that light socket is nevertheless live if you happen to
shove your finger in it when the switch is at the "on" position.
So if it is not leaking, what else is it doing?
7. Electricity is made up of two ingredients, negative and
positive. One ingredient travels along a wire covered with red
plastic, and the other along a wire covered with black plastic.
When these two wires meet together in what we call a plug, the
different ingredients are mixed together to form electricity.
Washing machines need stronger electricity, and for this a
booster ingredient is required. This travels along a wire covered
with green plastic.
8. Stronger electricity cannot be used for electric razors.
Electric razors make a fizzing sound when attached to a power
plug.
9. Electricity may be stored in batteries. Big batteries do not
necessarily hold more electricity than small batteries. In big
batteries the electricity is just shovelled in, while in small
batteries (transistors) it is packed flat.
10. Electricity is composed of small particles called electrons,
an electron weighing only I 1 837 as much as an atom of the
lightest chemical element, hydrogen, unless the Encyclopedia
Britannica is a liar.
Baffled!
Incurious people are content to take all this as read. They
press a switch and the light comes on, and that is all they know
about the miracle in their homes. This has never done for me. I
have to know how things work, and if I cannot find out from
some technical handbook— the Every Boys' Wonder Book series
does an advanced manual on electricity— then I combine such
information as I already have with simple logic. Thus it is very
easy to deduce that the light switch controls a small clamp or
vice which grips the wires very hard, so that the electricity cannot
get through. When the switch is flicked on the vice is relaxed
and the electricity travels to the light bulb where a bit of wire,
called the element, is left bare. Here, for the first time, we can
actually see the electricity, in the form of a small spark. This
spark is enlarged many hundreds of times by the curved bulb
which is made of magnifying glass.
Why, is our next question, do these light bulbs have a limited
life? As any schoolboy knows, heat converts oxygen into
moisture. When all the oxygen in the light bulb has become
liquified in this manner, it naturally quenches the electric spark.
Some years ago a man in Birmingham invented an everlasting
electric light bulb which, since it contained no oxygen, would
never go out. The rights in it were bought up by the Atlas people
who keep it locked in their safe.
Now we come to electricity as a source of power rather than
a source of light or heat. Why, when you plug in an electric
iron, docs it get hot, whereas when you plug in an electric fan
it does not get hot but whirrs round and round? The answer is
that when light or heat is required we use bare electricity, where-
as when power is required we keep the electricity covered up.
The constant flow of sparks, unable to escape, is converted into
energy. This energy is fed into a motor which makes things go
round and round.
I have not yet touched on fuse wire. It has always amazed me
that an industry which is so en/terprising in most respects— the
invention of colour electricity for use in traffic lights and the
harnessing of negative electricity for refrigeration are two
examples that come to mind— should still, two hundred years
after James Watt invented the electric kettle, be manufacturing
fuse wire too thin. I pass on a hint for what it is worth. There is
available from hardware shops a sturdy wire used mostly for
making chicken runs, and this is far more durable than the stufi
sold by electricians (who must, I appreciate, make a living). By
using chicken wire I now have a fuse box which— even when the
spin-dryer burst into flames due to too much booster electricity
having been fed into it — has for six months been as impregnable
as the Bank of England.
But why have fuse wire at all? I completely understand that
the fuse box is the junction at which the wires leading from the
power station join, or fuse with, the wires belonging to the
house, and that these two sets of wires have got to be connected
with each other somehow. But what is wrong with a simple
knot? Perhaps I might make this the subject of a paper for the
Electrical Journal which, I now see from the Writers' and Artists
Year Book, welcomes electro-technical contributions not
exceeding 3,000 words.
In some respects, I reiterate, my knowledge is imperfect. I
have not yet explored the field of neon signs — how do they make
the electricity move about? And the pop-up toaster — how does
it know when the toast is ready? With an electronic eye, pre-
sumably— and this brings us to another fruitful area. What is
the difference between electricity and electronics? Or is there a
difference? Is electronics now just the smart word to use, like
high-speed gas? How can an English computer speak French,
which requires a different voltage? Logic would answer these
questions too, and many of a more technical nature, but the
light over my desk has just gone out. A valve blown somewhere,
I expect.
Authors and Artists
NORMAN LEADER ALLEN
Norman Leader Allen, British physicist, was born
in 1927 and received his B.Sc. from the University
of Birmingham, England, in 1948 and his Ph.D. in
1951. Allen has been a staff member of Massachu-
setts Institute of Technology and is now a lec-
turer in the Electrical and Electronic Engineering
Department at the University of Leeds. In addition
to his book, Threshold Pressure for Arc Discharges,
he has written extensively in scientific journals on
arc discharges, cosmic rays and plasma physics.
ALBERT V. BAEZ
Albert V. Baez, born in Puebla, Mexico, in 1912,
received his B.A. at Drew (1933), an M.A. from
Syracuse (1936), and a doctorate in physics from
Stanford University (1950). He has taught at
Drew University, Wagner College, Stanford, and
Harvard. From 1949 to 1950 he was a physicist
in the aeronautical laboratory at Cornell, and
from 1951 to 1958 professor of physics at the
University of Redlands. He was physicist to the
Film Group of the Physical Science Study Com-
mittee, and for six years headed the science
teaching section at UNESCO in Paris.
STANLEY SUMNER BALLARD
Stanley S. Ballard, Professor of Physics and
chairman of the department at the University of
Florida, Gainesville, was born in Los Angeles
in 1908. He received his A.B. from Pomona
College, and M.A. and Ph.D. from the University
of California. He has taught at the University of
Hawaii, Tufts University, and has been a research
physicist at the Scripps Institution of Oceona-
graphy. Ballard has served as president of the
Optical Society of America. His specialities are
spectroscopy, optical and infrared instrumentation,
and properties of optical materials. Ballard is co-
author of Physics Principles.
JOHN M. CARROLL
John M. Carroll was born in Philadelphia in 1925,
and educated at Lehigh University, and Hofstra.
He was editor at Electronics Magazine from 1952
to 1964, became professor of industrial engineering
at Lehigh in 1964, and Associate Professor of the
Department of Computer Science, University of
Western Ontario, London, Ontario, Canada, since
1968. His professional work is in industrial engi-
neering and electronics.
charge of the first aircraft ground-controlled ap-
proach project. He has won the Kalinga Prize,
given by UNESCO for the popularization of science.
The feasibility of many of the current space devel-
opments was perceived and outlined by Clarke in
the 1930's. His science fiction novels include
Childhoods End and The City and the Stors.
ALBERT EINSTEIN
Albert Einstein, considered to be the most creative
physical scientist since Newton, was nevertheless
a humble and sometimes rather shy man. He was
born in Ulm, Germany, in 1879. He seemed to learn
so slowly that his parents feared that he might be
retarded. After graduating to the Polytechnic In-
stitute in Zurich, he became a junior official at
the Patent Office at Berne. At the age of twenty-
six, and quite unknown, he published three revo-
lutionary papers in theoretical physics in 1905.
The first paper extended Max Planck's ideas of
quantization of energy, and established the quan-
tum theory of radiation. For this work he received
the Nobel Prize for 1921. The second paper gave
a mathematical theory of Brownian motion, yield-
ing a calculation of the size of a molecule. His
third paper founded the special theory of relativity.
Einstein's later work centered on the general
theory of relativity. His work had a profound in-
fluence not only on physics, but also on philo-
sophy. An eloquent and widely beloved man,
Einstein took an active part in liberal and anti-
war movements. Fleeing from Nazi Germany, he
settled in the United States in 1933 at the Insti-
tute for Advanced Study in Princeton. He died
in 1955.
RICHARD PHILLIPS FEYNMAN
Richard Feynman was born in New York in 1918,
and graduated from the Massachusetts Institute of
Technology in 1939. He received his doctorate in
theoretical physics from Princeton in 1942, and
worked at Los Alamos during the Second World
War. From 1945 to 1951 he taught at Cornell, and
since 1951 has been Tolman Professor of Physics
at the California Institute of Technology. Professor
Feynman received the Albert Einstein Award in
1954, and in 1965 was named a Foreign Member
of the Royal Society. In 1966 he was awarded the
Nobel Prize in Physics, which he shared with
Shinichero Tomonaga and Julian Schwinger, for
work in quantum field theory.
ARTHUR C. CLARKE
Arthur C. Clarke, British scientist and writer is a
Fellow of the Royal Astronomical Society. During
World War II he served as technical officer in
LEOPOLD INFELD
Leopold Infeld, a co-worker with Albert Einstein
in general relativity theory, was born in 1898 in
Poland. After studying at the Cracow and Berlin
306
Universities, he became a Rockefeller Fellow at
Cambridge where he worked with Max Born in
electromagnetic theory, and then a member of the
Institute for Advanced Study at Princeton. For
eleven years he was Professor of Applied Mathe-
matics at the University of Toronto. He then re-
turned to Poland and became Professor of
Physics at the University of Warsaw and until his
death on 16 January 1968 he was Director of the
Theoretical Physics Institute at the university.
A member of the presidium of the Polish Academy
of Science, Infeld conducted research in theoretical
physics, especially relativity and quantum theories.
Infeld was the author of The New Field Theory,
The World in Modern Science, Quest, Albert
Einstein, and with Einstein The Evolution of
Physi cs.
K. SCOTT KINERSON
Dr. Kinerson was educated at the University of New
Hampshire, Rensselaer Polytechnic Institute, and
Michigan State University. After serving in the
U.S. Army from 1943 to 1946, he became Instructor
in Physics at the University of Massachusetts at
Fort Devens, in 1946. In 1948 he joined the staff
of Russell Sage College in Troy, New York as
Instructor in Physics. He is presently Chairman
of the Department of Physics and Mathematics at
that college. He is a co-author of Introduction to
Natural Sciences, Part I— The Physical Sciences,
l966\
THOMAS JEFFERSON
Thomas Jefferson, third President of the United
States, was born in 1743 at Shadwell in Goochland
County, Virginia. He studied Greek, Latin, and
mathematics at the College of William and Mary for
two years, and later became a lawyer. From 1768
to 1775 Jefferson was a member of the Virginia
House of Burgesses. In 1775 he was elected to the
Second Continental Congress, and in 1776 he drafted
the Declaration of Independence. Jefferson felt a
conflicting devotion to the tranquil pursuits of
science and public service. His interests ranged
over such fields as agriculture, meteorology, pale-
ontology, ethnology, botany, and medicine. He be-
lieved in the freedom of the scientific mind and the
importance of basing conclusions on observations
and experiment. Jefferson demanded utility of
science, hence his numerous inventions and interest
in improvements and simplifications of agricultural
tools and techniques, and in balloons, dry docks,
submarines, even the furniture in his home (swivel
chairs and music stands). Because of his promi-
nence as a public figure, he was influential in in-
creasing and improving science education in
America. He died on July 4, 1826, the fiftieth
anniversary of the Declaration of Independence.
MATTHEW JOSEPHSON
Matthew Josephson, prolific writer and magazine
editor, was born in Brooklyn in 1899. He received
his B.A. from Columbia University in 1920.
Josephson was successively editor of the Broom,
Transition, and The New Republ ic, which he left
in 1932. In 1948 he was elected to the National
Institute of Arts and Letters and olso was a
traveling Guggenheim fellow for creative literature.
He is the author of Zola and His Time, The Robber
Barons, and Portrait of the Artist as American.
ROBERT B. LEIGHTON
Robert B. Leighton, born in Detroit, Michigan in
1919, was first a student and then a faculty member
at California Institute of Technology. He is a mem-
ber of the International Astronomical Union, the
National Academy of Science and the American
Physics Society. Professor Leighton's work deals
with the theory of solids, cosmic rays, high energy
physics, and solar physics.
ABRAHAM S. LUCHINS
Dr. Luchins received a B.A. degree from Brooklyn
College (1935), M.A. degree from Columbia Univer*
sity (1936), and his PhD. at New York University
(1940). He was research assistant to the psycho-
logist Max Wertheimer, clinical psychologist in
the United States Army, and Director of Mental
Hygiene Clinic for the Veterans' Administration.
He was taught at McGill University, University of
Oregon, University of Miami, and since 1962 has
been Professor at the State University of New York
at Albany. His publications include: Logical Foun-
dations of Mathematics for Behavioral Scientists
(1965) and Group Therapy: A Guide (1964); and he
was a co-author of Introduction to Natural Science
(Parts I and II), 1968 and 1970.
DAVID KEITH CHALMERS MACDONALD
David Keith Chalmers MacDonald was born in
Glasgow, Scotland, in 1920 and received his M.A.
in mathematics and natural philosophy from Edin-
burgh University in 1941. After serving with the
Royal Mechanical and Electrical Engineers during
World War II, he received his Ph.D. in 1946 from
Edinburgh. Then he attended Oxford as a research
fellow and received a Ph.D. in 1949. In 1951 Dr.
MacDonald went to Canada and started a low tem-
perature physics research laboratory for the National
Research Council. MacDonald was appointed to the
physics department at Ottawa University in 1955
and elected Fellow of the Royal Society of London
in 1960. Aside from numerous articles in scientific
journals, he was the author of Near Zero: An Intro-
duction to Low Temperature Physics and Faraday,
Maxwell, and Kelvin. MacDonald died in 1963.
307
Authors and Artists
JAMES CLERK MAXWELL
See J. R. Newman's articles in Readers 3 and 4.
ALAN S. MELTZER
Alan S. Meltzer was born in New York in 1932 and
educated at the University of Syracuse, and at
Princeton, where he received his Ph.D. in astronomy,
in 1956. He was physicist at the Smithsonian Astro-
physical Observatory from 1956 to 1957. Presently he-
is Assistant Professor of Astronomy at Rensselaer
Polytechnic Institute at Troy, New York. His areas
of investigation include solar and stellar spectro-
scopy, and solar-terrestrial relations.
ALBERT ABRAHAM MICHELSON
Precision measurement in experimental physics
was the lifelong passion of A. A. Michelson (1852 —
1931), who became in 1907 the first American to
win a Nobel Prize in one of the sciences. Born in
Prussia but raised in California and Nevada,
Michelson attended the U.S. Naval Academy and
was teaching there in 1879 when he first improved
the methods of measuring the velocity of light on
earth. After a post-graduate education in Europe he
returned to the United States where he taught phy-
sics at the college that became Case Institute of
Technology, then at Clark University, and at the
University of Chicago. While in Europe he invented
the famous instrument called the Michelson inter-
ferometer and while in Cleveland at Case in 1887,
he and E.W. Morley improved this device in an
effort to measure the absolute velocity of the
Earth as it hurtles through space. The failure of
the Michelson-Morley aether-drift experiment was
an important result that showed a deep flaw in
19th-century physics. Although Michelson re-
mained a creative experimentalist in physical op-
tics, meteorology, astrophysics and spectroscopy
throughout his life, he died still believing in the
wave model of the nature of light and in his
"beloved aether." His experimental valu« of the
speed of light, refined still further just before his
death, remain the accepted value of one of the few
"absolute" constants in physics for several
decades.
JAMES ROY NEWMAN
James R. Newman, lawyer and mathematician, was
born in New York City in 1907. He received his
A.B. from the College of the City of New York and
LL.B. from Columbia. Admitted to the New York
bar in 1929, he practiced there for twelve years.
During World War II he served as chief intelli-
gence officer, U. S. Embassy, London, ond in
1945 as special assistant to the Senate Committee
on Atomic Energy. From 1956-57 he wassenior
editor of The New Republic, and since 1948 had
been a member of the board of editors for Scien-
tific American where he was responsible for the
book review section. At the same time he was a
visting lecturer at the Yale Law School. J. R.
Newman is the author of What is Science?,
Science and Sensibility, ond editor of Common
Sense of the Exact Sciences, The World of
Mathematics, and the Harper Encyclopedia of
Science. He died in 1966.
V. LAWRENCE PARSEGIAN
V. Lawrence Parsegian studied at M.I.T., Washing-
ton University, and New York University, obtaining
his Ph.D. in physics in 1948. He has been profes-
sor of nuclear science and engineering at Rens-
selaer Polytechnic Institute, since 1954, and
holds the distinguished Chair of Rensselaer pro-
fessorship. In addition to his research activities,
he has chaired a curriculum development project
to improve college science teaching.
PHYSICAL SCIENCE STUDY COMMITTEE (PSSC)
As one of the earliest curriculum development
groups, formed in 1956 and consisting of scientists
and educators, it produced materials for a new high
school physics course (first published in 1962).
These continue to be used by many students and
teachers in the U.S., and portions of the course
have been adapted also for use in other countries.
MATTHEW SANDS
Matthew Sands was born in Oxford, Massachusetts,
in 1919. He attended Clark College, Rice Institute
of Technology. During World War II he worked at the
the Los Alamos Scientific Laboratory. He was Pro-
fessor of Physics at the California Institute of
Technology before joining the linear accelerator
group at Stanford University. Professor Sands
specializes in electronic instrumentation for
nuclear physics, cosmic rays, and high-energy
physics. He served as chairman of the Commis-
sion on College Physics.
WILLIAM ASAHEL SHURCLIFF
Born in Boston in 1909, William A. Shurcliff was
educated at Harvard, receiving his Ph.D. in
physics in 1934. During the war he served as tech-
nical aide to the Office of Scientific Research and
Development, National Defense Research Committee,
and Manhattan project. Then he was with the Polar-
oid Corporation as senior scientist and project
leader. He is now a Research Fellow at the Elec-
tron Accelerator at Horvard. Shurcliff is the author
of Polarized Light: Production and Use and Bombs
308
at Bikini. His technical interests include emission
spectroscopy, absorption spectrophotometry, atomic
energy, gamma radiation dosimeters, microscope
design, and color vision. He has headed a citizen's
group to examine the deleterious effects of the
planned supersonic transport planes.
JAMES ALFRED VAN ALLEN
James Alfred Van Allen, discoverer of the "Van
Allen radiation belt," was born at Mt. Pleasant,
Iowa, in 1914. After his undergraduate work at
Iowa Wesleyan College, he received his M.S. and
in 1939 his Ph.D. from the State University of
Iowa, where he is now a Professor of Physics and
Astronomy. He has been a Carnegie research fel-
low, and a research associate at Princeton, and is
the recipient of numerous honorary doctorates. For
his distinguished work in nuclear physics, cosmic
rays and space probes, he has been awarded the
Hickman Medal from the American Rocket Society,
the Distinguished Civilian Service Medal of the
U.S. Army, and the Hill Award of the Institute of
Aerospace Science.
EDGAR VILLCHUR
Edgar Villchur is President and Director of Re-
search of the Foundation for Hearing Aid Re-
search in Woodstock, New York. He was born in
New York City in 1917 and received a M.S.Ed, from
the City College of New York. He has taught at
New York University, and was President and Chief
Designer of Accoustic Research, Inc., a manufac-
turing company in the high fidelity field.
GEORGE WALD
George Wald was born in New York in 1906 and re-
ceived his education at New York University and
Columbia University. He did research in biology at
the Universities of Berlin, Zurich, and Chicago,
and joined the faculty of Harvard University in
1935, where he now is professor of biology. He is
the recipient of many honors for his work on the
biochemistry of vision, including the Nobel Prize
in physiology and medicine for 1967. He is also
widely regarded as one of the outstanding teachers
of biology.
WILLIAM GREY WALTER
William Grey Walter was born in 1911 and received
his M.A. and Sc.D. (1947) from Cambridge Univer-
sity. He was a Rockefeller Fellow at the Maudsley
Hospital in England. W. Grey Walter is a pioneer in
the use of electroencephalography for translating
the minute electrical currents of the human brain
into physical patterns which may be studied for
the information they give us on brain processes.
Walter is the author of The Living Brain, Further
Outlook, The Curve of the Snowflake and articles
to various scientific journals.
309