Fritz Machiup & Una Mansfield - The Study of Information
Fritz Machiup & Una Mansfield - The Study of Information
THE STUDY OF
INFORMATION
INTERDISCIPLINARY MESSAGES
Edited by
Fritz Machiup
and
Una Mansfield
THE CONTRIBUTORS
Michael A. Arbib
Avron Barr
David Batty
Robert D. Beam
Toni Carbo Bearman
Robert C. Berwick
Thomas G. Bever
NUNC COCNOSCO EX PARTE
Margaret A. Boden
Kenneth E. Boulding
Walter Buckley
C. West Churchman
Richard H. Day
Ray C. Dougherty
THOMASJ. BATA LIBRARY
Murray Eden
TRENT UNIVERSITY
Peter Elias
Michael S. Cazzaniga
Saul Gorn
Douglas R. Hofstadter
Samuel Jay Keyser
Manfred Kochen
Richard N. Langlois
Fritz Machlup
Donald M. MacKay
Una Mansfield
Richard Mattessich
Mihajlo D. Mesarovic
George A. Miller
Elliott W. Montroll
Hassan Mortazavian
Joel Moses
Allen Newell
Charls Pearson
Alan J. Perlis
Zenon W. Pylyshyn
W. Boyd Rayward
Jesse FT Shera
Vladimir Slamecka
Myron Tribus
Peter Wegner
Patrick Wilson
Vladimir Zwass
Digitized by the Internet Archive
in 2019 with funding from
Kahle/Austin Foundation
https://archive.org/details/studyofinformatiOOOOunse
THE STUDY OF
INFORMATION
THE STUDY OF
INFORMATION
Interdisciplinary Messages
Edited by
with a Foreword by
GEORGE A. MILLER
A Wiley-lnterscience Publication
“A Wiley-Interscience publication.”
Bibliography: p.
Includes index.
1. Information science—Addresses, essays, lectures.
I. Machlup, Fritz, 1902-1983. II. Mansfield, Una.
10 987654321
CONTRIBUTORS
[Editors' Note: The following list does not include past affiliations, honors, and awards or major
publications of contributors. Many of these publications are included in the Cumulative List of
References. Our main objective is to show the major disciplinary commitments of the authors;
we therefore include their past and present offices (but not council memberships) in national
professional associations, and editorial duties and editorial-board memberships for learned
journals and professional periodicals.]
IX
X
FOREWORD
that was published in 1962 by the Princeton University Press: The Produc¬
tion and Distribution of Knowledge in the United States. The two longest
chapters were IV, entitled Education, and V, Research and Development.
The book was immediately recognized as a significant and pioneering
work, the first serious discussion of a basic change in American life. In the
concluding chapter Machlup stated the clear implication: “If employment
opportunities continue to improve for high-level-knowledge-producing labor
and to worsen for unskilled manual labor, the danger of increasing unem¬
ployment among the latter becomes more serious” (p. 397). Needless to say,
the trend has continued: his estimate that knowledge production and distri¬
bution accounted for 29 percent of the adjusted gross national product in
1958 was subsequently updated by Marc Uri Porat to 46 percent in 1967. And
the unemployment of unskilled labor has indeed become more serious.
In 1971, when he retired as Walker Professor of Economics and Interna¬
tional Finance at Princeton, Machlup decided that he had a moral obligation
to update his statistical analysis. In addition to his duties as professor at
New York University and a full agenda of other research, he undertook an
eightfold expansion of the 1962 book: the initial plan was for a series of eight
volumes, roughly one volume for each chapter of the earlier book, the whole
to be entitled Knowledge: Its Creation, Distribution, and Economic
Significance, with the Princeton University Press as publisher.
Volume I, Knowledge and Knowledge Production, was published in 1980.
According to the original plan, the second volume was to survey the
branches of learning and the information sciences, and to analyze the eco¬
nomic concept of human capital. As the work progressed, however, it grew
into three separate books, expanding the planned series from eight to ten
volumes. The first part, The Branches of Learning, appeared as Volume II in
1982. Volume III, The Economics of Information and Human Capital, was
completed only weeks before Machlup died of a heart attack on January 30,
1983, at 80 years of age.
Volume IV was to deal with the information sciences, the remaining part
of the originally planned second volume. In preparation, Machlup persuaded
39 information scientists to write a total of 56 essays on their various special¬
ties so that he, in his role as an editor, could go to school under the experts—
could “see the stir of the great Babel, and not feel the crowd.” The result is
the present book, prepared in collaboration with Una Mansfield. It is not
Volume IV, of course, but some indications of Machlup’s reactions to this
heroic exercise can be gathered from the co-authored Prologue and his own
Epilogue.
Probably nothing less than a Grand Design could have persuaded anyone,
even the enormously energetic Machlup, to organize such an enterprise as
this. The information sciences are highly diverse. Someone who speaks of
information in his own context may know of someone else who uses it in a
different context, yet never find the occasion or the impulse to explore the
similarities and differences. Now Machlup has made that exploration for us,
FOREWORD XI
and made it with a sensitivity to semantics and logic that few could match.
Discourse across these disciplinary boundaries will surely proceed more
smoothly in the future. Experts can join with The Celebrated Intelligent
Layman in their gratitude for the resulting improvement in mutual intelligi¬
bility. The topic is important, sufficiently important that, even if we cannot
speak in unison, we should speak as clearly as possible.
Kenneth Boulding once commented that Machlup had what he called the
anti-Midas touch—even if he were to touch gold, he would turn it to life. All
the contributors to this volume, all who endured his compulsive criticism
and enjoyed his appreciative praise as he tried to bring our essays to life, join
with Una Mansfield and me in dedicating our efforts here to the memory of
our charming friend and mentor, Fritz Machlup.
George A. Miller
xiii
XIV PREFACE
Nine lead papers and between three and five discussion papers for each of
the nine were commissioned. The writers of subsidiary papers were asked to
offer critical comments, elaborations, or supplementary observations on the
subject of the lead paper. The author of the lead paper would then come
back with a rejoinder. This procedure was designed to produce a discussion
as lively as a round-table debate and yet as well-thought-out as an argument
presented for publication in a learned journal or collective volume. Al¬
together, 56 contributions from 39 authors, plus an analytical Prologue and
Epilogue by the co-editors, resulted from this undertaking. All responses
and rejoinders were given titles descriptive of the messages delivered, and
the work was divided into nine sections dealing with cognitive science,
computer and information science, artificial intelligence, linguistics, library
and information sciences, cybernetics, information theory, mathematical
system theory, and general system theory. The format of lead paper/
comments/rejoinder worked well, except in Section 3 where the four papers
were written independently, with “endnotes” rather than a rejoinder by the
lead author, and in Section 4, which has no rejoinder.
The interdisciplinary character of the project is evident, whether authors
are listed by departmental affiliation or by professed specialty. As a matter
of fact, diversity of opinions, especially in matters of methodology, is not
confined to diversity of research commitment or of departmental affiliation;
disagreements are most conspicuous, and aired with greatest vehemence,
among those whose research interests are most alike. All in all, however,
more can be learned from the interplay of conflicting ideas, from the argu¬
ments and counterarguments on each issue, than from the best-formulated
but monolithic expositions of the fields.
In view of our interdisciplinary readership, a few editorial conventions
had to be imposed: for instance, technical “jargon” was discouraged and
mathematical notation reduced to the necessary minimum. First names were
used in the text to introduce scholars who though well known in some
disciplines might not be known in others. In citations and quotations, initials
PREFACE xv
of authors with common family names were included for easy reference. The
manuscript was virtually complete at the time of Professor Machlup’s death,
with the exception of his own Epilogue, the Cumulative List of References,
and the Index. The parts missing from the Epilogue have been indicated by
me in an Editor’s Note at the point in the text where his outline had placed
them; the parts that have been published are exactly as Machlup had written
them.
Apart from Fritz Machlup, the information community lost two other
outstanding scholars during the course of this research: Colin Cherry died in
November 1979, about the time Machlup was inviting him to act as an
adviser to the project; and Jesse H. Shera died in March 1982, just a few
months after he had submitted his paper in Section 5. Each has left us a rich
legacy in his writings.
A project of this magnitude involves the cooperation of many organiza¬
tions and individuals. Funding for the research was provided by grants from
the National Science Foundation to New York University, and from the
Spencer Foundation and the Earhart Foundation to Princeton University.
Our thanks to these foundations and institutions for their support. Howard
L. Resnikoff and his successor as director of the NSF Division of Informa¬
tion Science and Technology, Edward C. Weiss, provided encouragement
and guidance to Professor Machlup; and Herbert S. Bailey, Jr., director of
the Princeton University Press, and Robert F. Rich, editor of the journal
Knowledge, were trusted friends whose advice helped shape his plans.
Scholars who served as advisers to the project and nominated writers of
papers and comments included Russell Ackoff, Saul Gorn, Kenneth A.
Klivington, Donald E. Knuth, Lewis Levine, J. C. R. Licklider, Mitchell
Marcus, Mihajlo Mesarovic, George A. Miller, Alan J. Perlis, Claude E.
Shannon, Herbert A. Simon, David Slepian, Don R. Swanson, and Terry
Winograd. Four of them—Gorn, Mesarovic, Miller, and Perlis—contributed
papers of their own, as did the 35 other scholars in the List of Contributors.
We are grateful to all of them, especially the authors of lead papers who
were willing to have their work “critiqued” so intensively.
Professor Machlup benefited from discussing parts of our Prologue and
his Epilogue with several colleagues and family members, among them John
Woodland (Woody) Hastings, Richard C. Jeffrey, Richard N. Langlois,
Stefan Machlup, George A. Miller, Allen Newell, and Kenneth Stigleitz.
He would also want me to express his gratitude to Mary Taylor Huber and
Peter Swire, who provided valuable research assistance in the earliest phase
of the project, and to Rosa Schupbach for secretarial support throughout its
duration. Eliana Covacich and Robert Korchak at New York University
were especially helpful to him; and Michael McGill and Helene Ebenfield at
NSF showed understanding as program and project officials. As chairman of
the economics department, Stephen M. Goldfeld acted as “host” for Profes¬
sor Machlup’s Princeton-based research during his lifetime and bore respon¬
sibility for its successful completion after his death. Two good friends of
XVI PREFACE
Una Mansfield
Princeton University
October 1983
ACKNOWLEDGMENTS
The authors and editors are indebted to the publishers for their permission to
reprint material from copyrighted works, as follows:
In Section 1
In Section 4
In Section 6
XVII
XVIII
ACKNOWLEDGMENTS
In Section 7
Tribus, Myron, “Thirty Years of Information Theory,” in Levine,
Raphael D., and Tribus, Myron, eds., The Maximum Entropy Formal¬
ism Conference, MIT, 1978, copyright © by the Massachusetts Insti¬
tute of Technology; paper reprinted by permission of the MIT Press.
CONTENTS
PROLOGUE
SECTION 1
SECTION 2
xix
XX CONTENTS
SECTION 3
SECTION 4
Rules and Principles in Phonology and Syntax, Samuel Jay Keyser 327
SECTION 5
SECTION 6
Cybernetics 409
Murray Eden
Cybernetics: Past and Present, East and West, Peter Elias 441
SECTION 7
SECTION 8
SECTION 9
EPILOGUE
INFORMATION
PROLOGUE
CULTURAL DIVERSITY IN
STUDIES OF INFORMATION
Fritz Machiup and Una Mansfield
The idea of two cultures in the academic world has been most forcefully
presented by C. P. Snow. According to him, a deep intellectual gulf divides
the mathematically minded laboratory-dwellers, engaged in the natural sci¬
ences, from the book-loving denizens of the library stacks, the literary intel¬
lectuals. [Snow, 1959.] The two cultures, sharing a mutual incomprehension
and disrespect of each other, are what now (in violation of semantic tradi¬
tion) are called science and the humanities. [Machiup, 1980, pp. 62-84.]
Snow later became aware of the existence of other fields. He conceded that
recognition of the social sciences would present us with “something like a
third culture,” but he denied the good sense of considering all possible
branches and twigs of learning that might give us “two thousand and two”
cultures. (Snow, 1964, pp. 66, 70.) We have chosen to deal here with only
thirty or forty cultures, a limitation explained by the fact that we are not
going to deal with the entire universe of learning but with only that part of it
that is characterized by the keyword information.
THE PROJECT
Our project is to analyze the logical (or methodological) and pragmatic rela¬
tions among the disciplines and subject areas that are centered on informa¬
tion. Disciplines (sciences, academic areas of research and teaching) are
orderly arrangements (metaphorically called bodies) of coherent thoughts,
formulated as propositions, about things (sense-objects or thought-objects)
deemed worthy of being known (i.e., being believed with some degree of
confidence) and being passed on. In other words, disciplines are what a
number of people, respected for having read widely and for being read by
other widely read people, have claimed to be disciplines. (We shall return
later to the question of how this pronouncement can be made operational.)
Relations among different disciplines are therefore relations among the ex¬
pressed thoughts of selected scholars (scientists). Logical relations among
3
4 FRITZ MACHLUP AND UNA MANSFIELD
Interdisciplinary Explorations
much more than the name. If we have failed and are still at sea, it may be our
fault: Explorers do not always succeed in learning the language of the na¬
tives and their habits of thought.
One scholar who is at home in several of the different areas of study—
Herbert Simon, spiritus motor in many scientific endeavors—likened our
project to an anthropological exploration: We go into areas whose inhabi¬
tants speak foreign tongues (with many words sounding like words in our
own language but having very different meanings); we try to find some
guides to help us learn the meanings of these strange sounds; and we try to
make sense of what we see and hear, yet we probably misunderstand much
and are bewildered by even much more. Another scholar refused to compare
our work to anthropological research, chiefly because anthropologists often
have taken vows of intellectual chastity, so that they can embark on their
explorations untroubled by prior knowledge, immaculately free from pre¬
conceptions, and innocent of philosophical liaisons. What is really needed
for a successful inquiry is “an interdisciplinary epistemology . . . integrated
with philosophical understanding and with psychological and biological
knowledge.” We tend to agree with this judgment expressed by Margaret
Boden but should quickly admit that the statement, made in her paper in this
volume, was meant to refer chiefly to computational insights. We find it
equally valid for all interdisciplinary inquiries in the fields concerned with
information.
The intellectual requirement to break down the barriers between disci¬
plines has long been recognized. In virtually all branches of learning, we can
observe an ongoing conflict between specialization and integration, separat¬
ism and unificationism, isolationism and scholarly cosmopolitanism, or, in
brief, fission and fusion. The warning that we ought not to become
Fachidioten—the German pejorative word for excessively specialized ex¬
perts ignorant of things outside their narrow fields—has been sounded in all
areas of knowledge, but it is noteworthy that some of the founders of the
disciplines of information have been particularly concerned about the schol¬
arly seclusion of their fellow researchers.
Although a call for “mere intellectual communication across the bound¬
aries of these several disciplines is not enough”—to quote again from
Boden’s paper—listening to interdisciplinary messages is better than plug¬
ging one’s ears. International and interdisciplinary symposia were organized
soon after Norbert Wiener came out with his Cybernetics [1948] and Claude
Shannon with his paper on “A Mathematical Theory of Communication”
[1948]. The First International Symposium on Information Theory was held
in London in the summer of 1950. Among the participants were “mathema¬
ticians, physicists, engineers, linguists, physiologists, geneticists—in fact
folk from almost every branch of science.” (MacKay, 1969, p. 9.) All
branches of learning mentioned in this quotation are traditional disciplines,
represented in the conventional structure of every university. When similar
interdisciplinary symposia are held nowadays—32 years after that first
FRITZ MACHLUP AND UNA MANSFIELD
6
1 We limited ourselves in the preceding to the three best-known designations but may add here
another six of lesser currency: informantics, informatistics, informatology, informology, infor-
metrics, and infometrics. We are indebted to Alvin M. Schrader, who not only compiled a long
list of designations but also knows the names of the terminologists responsible for coining them.
[Schrader, 1984.]
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 7
fields of scientific work . . . which have been explored from the different sides
of pure mathematics, statistics, electrical engineering, and neurophysiology, in
8 FRITZ MACHLUP AND UNA MANSFIELD
which every single notion receives a separate name from each group, and in
which important work has been triplicated or quadruplicated, while still other
important work is delayed by the unavailability in one field of results that may
have already become classical in the next field.
It is these boundary regions of science which offer the richest opportunities to
the qualified investigator. (Wiener, 1948, p. 2.)
In the same vein, Ludwig von Bertalanffy, the biologist and one of the
pioneers of general system theory, wrote that
Splitting, or fission, of academic disciplines has been observed for over two
thousand years. Chemistry, for example, was an early spin-off from physics,
both being exhibited as separate sciences in philosophical discussions and
classifications of natural philosophy. After various subdivisions had oc¬
curred, a merger of parts of physics and chemistry proved expedient: It
resulted in physical chemistry. Several splits and mergers—fissions and
fusions—have occurred in the life sciences, with biology being divided into
several subdisciplines and merged with parts of physics and chemistry. This
process led to biophysics and biochemistry, with the latter splitting again
into chemical biochemistry and biological biochemistry. [Machlup, 1982, pp.
101, 152.] Similar sequences of splits and mergers have occurred in moral
10 FRITZ MACHLUP AND UNA MANSFIELD
being read widely by respected scholars who have read widely and are being
read widely by respected . . . and so forth—though the circle of scholars
need not be large. There are operational tests for the existence of such a
consensus among recognized scholars. The tests include such things as ref¬
erences to certain keywords characteristic of the new discipline (or spe¬
cialty) in subject indexes of recent textbooks and in titles of articles and
books; the appearance of survey articles summarizing the contents of publi¬
cations in the new field; the extension of the classification scheme of a larger
discipline in order to provide one or more special headings for listing publi¬
cations in the new discipline (or specialty); the appearance of books of
readings and volumes of conference proceedings focusing on the new disci¬
pline (or specialty); and, finally, courses offering instruction in the new
subject in universities, first to graduate students, later, perhaps, also to
undergraduates.3
A novel approach to the identification of specialties that are presumably
“ready to become recognized as having arrived” relies on citation analysis.
The analyst finds clusters of articles and books linked either by cocitation or
by bibliographic coupling. In cocitation, different earlier publications are
linked by being cited together in the same new pieces; in bibliographic
coupling, different new publications are linked as they cite the same earlier
pieces. [Garfield, Malin, and Small, 1978, p. 185.] A similar approach can
establish mutual-citation societies by compiling a citation matrix: In a new
or emerging field, whom did the writers cite and by whom were they cited?
If, however, a small group of people cite one another but cite hardly anyone
else, must one not suspect the existence of a club, a cult, or a conspiracy?
Perhaps so; yet one should not on this ground deny the group recognition as
a bona fide field of study, especially not if some of its members are scholars
also respected outside the group.
3These tests have actually been carried out to ascertain that the economics of knowledge and
information has become an established specialty within the field of economics. [Machlup, 1984,
chap. 9.] The number of titles published in this specialty exceeds 20,000.
12 FRITZ MACHLUP AND UNA MANSFIELD
We have discouraged some of our contributors from using much space for
discussing whether a particular discipline is true science or something else.
Several of them decided, even without our advice, not to expatiate on this
sterile topic. We did not, however, wield a red pencil to delete all comments
on whether something was a science, or “only” philosophy, speculation,
technology, practical art, professional training, and so forth. We knew that
many writers in information science had guilt feelings about the fact that this
discipline had neither discovered new laws nor invented new theories and
therefore did not deserve recognition as a science. Such an inferiority com¬
plex is the result of indoctrination with an outmoded philosophy of science,
with persuasive (propagandist) definitions of science and scientific method.5
The restrictive meaning of science, taught in courses and texts in experi¬
mental natural sciences and designed to exclude other academic disciplines
from the class of sciences, occurs only in English. Neither French nor Ger¬
man—nor any other language, to our knowledge—has words to express the
narrow concept of science or any expressions equivalent to scientific
method. [Machlup, 1980, pp. 67-69.] If the restriction of the term to eligible
disciplines could serve any scientific purpose, the French, Germans, Rus¬
sians, Japanese, and others would surely have coined words to express what
their English-speaking colleagues had in mind. Still, some contributors to
this volume do allude to conflicts between science and philosophy, science
and unverified belief, science and art, science and technology, science and
profession. These issues are therefore briefly addressed in the Epilogue by
Fritz Machlup.
For present purposes, we may confine ourselves to the statement that we
do not care, and no one else need care, whether information science, library
science, computer science, decision science, system science, or any of the
disciplines discussed in this volume, are genuine sciences. Fittle depends on
the decision to deny this honorific designation to a particular discipline.6 If
little or nothing follows from the distinction between science and nonsci¬
ence, it is not worth investing much time making or defending it. Too bad
that even a little time has to be spent showing that the distinction is mis¬
chievous.
The titles of the papers contributed to this volume contain the names of
twenty disciplines or subdisciplines. The texts of the papers contain refer-
5“. . . the attempt to define scientific method or to analyze science is a search for a persuasive
definition . . . the term ‘science’ has no definite and unambiguous application. (Black, 1949, p.
69.) “Neither observation, nor generalization, nor the hypothetic-deductive use of assump¬
tions, nor measurement, nor the use of instruments, nor mathematical construction nor all of
them together—can be regarded as essential to science.’’ (Ibid., pp. 80-81.)
6If someone points to the differences in salaries of librarians, physicists, and computer special¬
ists, the explanation lies in supply and demand, not in discriminatory designations of their
fields. However, it may be possible that agencies and foundations supporting research are
persuaded to favor grant applications for supposedly more scientific projects: The charge of
being ‘unscientific’ is not mere namecalling; it is a charge with financial consequences. It has to
do with what research gets funded and therefore with what research gets carried out. (Lakoff,
1978, pp. 267-268.) We suspect, however, that it is not the magic of the word scientific that
charms the grantors but rather the lure of research leading to practical findings.
14 FRITZ MACHLUP AND UNA MANSFIELD
7 The issue of practical versus pure sciences is perhaps most easily clarified by contrasting
medical science with medical practice. Clearly, medical research, teaching medicine, a.n& prac¬
ticing medicine are three different activities. No matter whether students of medicine plan to
become researchers, professors, or practitioners, they will have to be educated in several
medical and cognate sciences—anatomy, physiology, chemistry, biochemistry, pharmacology,
cardiology, ophthalmology, otolaryngology, gynecology, psychiatry, and several others. The
majority of medical students plan to take a degree of doctor of medicine and obtain a license to
practice medicine. In their capacity as practitioners, doctors are not scientists, though they
have mastered a good many sciences. Those engaged in medical research and/or in teaching
medicine may spend part of their time as practicing physicians; their scientific work consists of
research, publication, and teaching; but many medical doctors double as scientists and prac¬
titioners. None of the sciences taught in medical schools is less of a science due to the fact that
the ratio of those who have studied it and are “only” practitioners is much larger than in other
fields. An enthusiastic teacher-scholar may take pride in the fact that a larger percentage of
students choose to engage in research rather than in medical practice, but this does not reflect
on the discipline professed. Likewise, a medical school that produces relatively large numbers
of medical researchers will have a much better reputation than one that produces only prac¬
titioners. But this has nothing to do with the scientific character of disciplines taught in medical
schools. It is possible, of course, that the way the discipline is presented to students is affected
by the knowledge of the vocational-career orientation of most students, and, thus, the scientific
level of the presentation might be lowered.
16 FRITZ MACHLUP AND UNA MANSFIELD
Semiotics and linguistics, the science of signs and the science of those
signs that make up language, respectively, deal with observable objects.
Although a good deal of the reasoning in these disciplines is formal and,
thus, in the domain of mental construction, the subject matter is no doubt
empirical.
We now come to disciplines of a rather different character, and we want
to state that we have approached them with a minimum of preconceptions
regarding their methodological status. After careful study of the papers in
this volume and of some of the earlier literature on cybernetics and general
system theory (including mathematical system theory), we have come to the
following conclusions about the relation between these fields and their
places in the universe of learning.
Cybernetics is a method of examining interactive parts of any system of
communication and control, with particular emphasis on feedback and
homeostasis. It is part and parcel of general system theory, which is con¬
cerned with these and also other relations among parts of a whole and
between the whole and its environment. General system theory, which thus
comprises cybernetics, is a method of organizing one’s thinking in terms of
interrelated elements in closed or open sets. It is part of the general
methodology of nomological disciplines. General methodology, which thus
comprises general system theory, is “the second part of logic” (sometimes
called material logic in contrast to formal logic). (Windelband, 1913, p. 22.)
To summarize, cybernetics is a part of general system theory, which is a
part of methodology, which is a part of logic. All of these may be given
the designation of disciplines or sciences, though not empirical sciences.
Whenever formal sciences are applied to the interpretation of phenomena,
or of relations among phenomena, the resulting propositions are parts of the
discipline or disciplines that are concerned with the phenomena or relations
in question. To offer examples, mathematics and methodology applied to
molecular biology yield molecular biology; system theory applied to macro¬
economics yields macroeconomics; and cybernetics applied to neuro¬
physiology yields neurophysiology.
More will be said later, both about cybernetics and general system theory
in special sections of this introductory essay. Our next question for the
present discussion of formal versus empirical science concerns the status of
information theory in the narrow form of the mathematical theory of com¬
munication. With its progenitor, Claude Shannon, and his precursors, Harry
Nyquist and Ralph Hartley, all of them researchers at Bell Laboratories
doing mission-oriented research on electrical circuitry and signal transmis¬
sion, one would hardly have any serious doubts about this kind of work
being empirical. Yet, we shall argue that the strategic propositions of Shan¬
non’s theory are purely formal uses of general probability theory applied to
assumed states of assumed objects; they are analytical relations derived
from definitions and other resolutions.
This statement may sound strange to ears not attuned to methodological
18 FRITZ MACHLUP AND UNA MANSFIELD
Information Science
We have detected at least four main uses of the term information science in
the literature: (1) In its broadest sense, it stands for the systematic study of
information and may include all or any combination of the academic disci¬
plines discussed in this volume; (2) when included in the phrase computer
and information science, information science denotes the study of the phe¬
nomena of interest to those who deal with computers as processors of infor¬
mation; (3) in library and information science, it indicates a concern with the
application of new tasks and new technology to the traditional practices of
librarianship; and (4) in its narrow sense, information science is used as the
name for a new area of study that is evolving from the intersection of the
other three mentioned areas, with perhaps a special interest in improved
communication of scientific and technological information and in the appli¬
cation of well-tested research methods to the study of information systems
and services.
The earliest formal use of the term information science in the United
States seems to have been in the description of a program in computer and
information science at the Moore School of Electrical Engineering, Univer-
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 19
The early work in the broad area of information science is well characterized
by the subheading chosen for these paragraphs. Scientists and scholars from
a variety of disciplines were attracted to the new area of research and teach¬
ing; they brought with them their own research methods and terminology
resulting in a Babel of sorts. Librarians found it difficult to interact with
mathematicians; cognitive psychologists had problems in dealing with physi¬
cists; and social epistemologists considered engineers to be a breed apart.
This variety of interests is reflected even today in the membership of the
only society in the United States that seeks to cater to all information profes¬
sionals—The American Society for Information Science. [King, Krauser,
and Sague, 1980.]
The great melting pot of information science included, among other
things, electrical engineering (e.g., research on signal transmission over noisy
channels), computer technology (e.g., research on information-processing
by machines), biological sciences (e.g., research on information-processing
in living systems), behavioral sciences (e.g., research on cognitive pro¬
cesses), and social sciences (e.g., research on the sociology of knowledge).
Not that there were many who engaged in all of these areas of inquiry, but
there were some who concentrated on one or two and did so under the
banner of information science. However, assembling under a banner is not
the same thing as eliminating cultural diversities in the processes ascribed to
the melting pot: The diversities remain and no unified science emerges.
We submit that most of the confusion caused by the use of the term informa¬
tion science in its broadest sense could be avoided by the addition of the
plural 5. The information sciences could then take their place alongside the
natural sciences, the social sciences, and other umbrella terms that indicate
a grouping of disciplines and fields of study that share a common characteris¬
tic. The bond among the information sciences is, of course, their focus on
information as the object of study, though it is important to bear in mind that
the word information is interpreted very differently by various groups of
20 FRITZ MACHLUP AND UNA MANSFIELD
What has been said about missing hyphens and the plural t holds also for the
combination between the study of the library and the study of information.
8Most schools and even some universities offer courses called science and social science,
evidently presenting surveys of several disciplines under each title. We consider the omission of
the plural s in these instances as a benighted practice.
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 21
One needs to decide whether two separate but cognate subjects ought to be
housed under one roof or whether they should be fused into one science with
the research effort fully integrated. Most writers and teachers have not made
up their minds, and by omitting both hyphens and the plural y, they put an
unstructured string of words on their letterheads.
The type of information science that grew out of the documentation
movement of the 1960s, and is taught mainly in library schools, has a focus
different from that of the computer scientists. The emphasis is on better
techniques for managing recorded information in whatever medium it re¬
sides. There is concern with how the records will be used and, therefore,
with their intellectual content. Much of the information-science activity in
this field centers on the application of new technology to traditional library
functions of acquiring, storing, retrieving, displaying, and disseminating
records—in other words, on technically improved librarianship. Thus far,
there is little evidence that the new technology has been exploited to pro¬
duce radically new ways of handling the store of knowledge recorded in
library and other materials. But there is no doubt that new tasks have been
assigned to librarians, for example, the management of information centers
concerned with nonbibliographic information, participation in resource¬
sharing networks made necessary by the increasing volume of recorded
knowledge and dwindling library budgets, retrieval from electronic online
databases made possible by the new computer and communications technol¬
ogy, and so on.
The eagerness with which library schools have moved to incorporate the
word information into their titles is proof that their deans and faculties view
the new technological developments in information handling as vital for their
growth or even their survival. A recent article pointed out that “ ‘Informa¬
tion’ is now contained in the names of fully 37 of the 70 schools appearing in
the most recently revised list of programs accredited by the American Li¬
brary Association, although only one (Syracuse’s School of Information
Studies) has forsaken the word ‘library’ altogether.” (Harter, 1982, p. 40.)
But although library schools are quick to enlarge their curricula to include
aspects of information studies, there is a question whether this is always a
substantive change in disciplinary orientation. Often courses in information
science consist merely of teaching students to use a new tool, the com¬
puter—something that will eventually be done at earlier stages in the school¬
ing process, at the secondary, if not elementary, level. (Learning to use a
computer will soon be regarded as a basic skill.) On a somewhat higher level
is teaching data retrieval from online electronic databases, which is now
taught and may well remain in the curricula of library schools. Still, this sort
of training cannot reasonably be regarded as information science. To qualify
for that designation, the respective studies should involve designing, build¬
ing, and programming the online retrieval systems.
The discipline of library science, or librarianship, onto which this new
type of information science is being grafted has a long and proud tradition of
service. Its origins and development, and the disciplinary dilemma it faces in
22 FRITZ MACHLUP AND UNA MANSFIELD
9For a discussion of the concept of the information professional, and the results of a statistical
survey of information professionals employed in organizations in the United States, see The
Information Professional: Survey of an Emerging Field. [Debons, King, Mansfield, and Shirey,
1981.]
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 23
the area make significant progress, they will adopt identifying titles of their
own, as did the area of artificial intelligence within computer science.
Computer Science
Our first question about computer science is whether its subject is the com¬
puter itself, a highly complicated machine, or rather what is being done with
computers, namely, the processing of all sorts of information. This is contro¬
versial, even among those who teach computer science, with those who
favor the information focus recommending that the name of the discipline be
computer and information science or, better still, following the continental
European practice, informatics. [See Gorn’s lead paper in this volume.]
In their preliminary recommendations for an undergraduate program in
computer science, the Curriculum Committee of the Association for Com¬
puting Machinery supported the second position:
Computer Science is concerned with information in much the same sense that
physics is concerned with energy; it is devoted to the representation, storage,
manipulation and presentation of information in an environment permitting
automatic information systems. As physics uses energy transforming devices,
computer science uses information transforming devices. (ACM Curriculum
Committee, 1965, p. 544.)
The first position has been taken by Anthony Ralston. He insisted that
“information is no more uniquely the province of computer science than
energy is of physics. . . . The unique aspect of computer science is the
computer.” (Ralston, 1971, p. 1.) Many authoritative writers in the field
concur with this view: “Phenomena breed sciences. There are computers.
Ergo, computer science is the study of computers.” (Newell, Perlis, and
Simon, 1967, p. 1373.)
The machine in question is the general purpose automatic electronic digi¬
tal computer. Digital computers deal with discrete (numeric and nonnu¬
meric) quantities, as opposed to analogue computers, which deal with con¬
tinuous flows.10 Design of both types of computers was a specialty of the
field of electrical engineering, and operating and maintaining analogue com¬
puters has remained a subspecialty within that discipline. However, the
enormously complex task of designing, programming, operating, and main¬
taining the all-purpose digital computer gave rise to the development of a
separate discipline, computer science.
The choice of the name computer for the machine in question has proved
to be somewhat misleading. To be sure, the initial applications of these
machines were to processing numbers. Digital computers, however, can and
do manipulate entities that are symbolic representations of other things. We
have no estimate, but no doubt a very large portion of their use now is for
processing nonnumeric contents (e.g., texts, images, and graphics); there¬
fore, we agree with those who have suggested that symbol manipulator
would have been a more appropriate name for the machine. [For example,
Ralston, 1971.]
mathematics department; the labor of synthesis being too great for the insight
into theory each such study provides. Yet these algorithms . . . deliver the
power of the computer to the most demanding applications of our technological
society. (Perlis, 1968, p. 71.)
Perhaps we should leave the last word on this subject with Herbert
Simon, who has said that the “highly abstractive quality of computers
makes it easy to introduce mathematics into the study of their theory—and
has led some to the erroneous conclusion that, as a computer science
emerges, it will necessarily be a mathematical rather than an empirical
science.” (Simon, 1969 and 1981, pp. 18-19.)
There is little doubt about the engineering nature of computer science: It
designs and builds artifacts for all to see. The structure of a computer system
consists of hardware (physical components) and software (specialized pro¬
grams for managing the hardware). Hardware comprises electronic, electro¬
mechanical, and other devices for switching, storing, and communicating
signs and signals (in the form of electrical impulses) within the system.
Software consists of “a class of computer programs . . . which are used to
aid in the production, debugging, maintenance, and orderly running of the
other programs.” 12 It may be “more important than computer hardware in
determining the productivity of computer installations.” (Rosen, 1976, p.
1283.) The production of software for computers has become such a critical
and costly activity that in recent times attempts have been made to apply
engineering principles to it, with a view to standardization. This has led to
the evolution of a new specialty within computer science called software
engineering. [Wegner in this volume.]
Closely related to engineering is the question of design. In his widely read
monograph, The Sciences of the Artificial, Herbert Simon argues that “a
science of design not only is possible but is actually emerging at the present
time. It has begun to penetrate the engineering schools, particularly through
programs in computer science . . . .” (Simon, 1969 and 1981, p. 58.) Saul
Amarel points to a “fundamental reason for a close coupling between com¬
puter science and a science of design. It comes from . . . concern . . . with
the information processes of problem solving and goal-directed decision
making, which are at the core of design.” (Amarel, 1976, p. 314.)
12The layman often has difficulty in distinguishing between the hardware, firmware, software,
and programs of computer systems. One measure of distinction among them is their degree of
modifiability. Hardware refers to the material parts of the computer system; it cannot be
modified by a user. Firmware refers to computer programs (sets of instructions) that have been
embodied in a physical device that can form part of a computing machine; it also cannot be
modified by a user. Software refers to various programming aids that facilitate a user’s efficient
operation of the computer equipment (examples are assemblers, generators, subroutine librar¬
ies, compilers, operating systems, etc.); it is subject to some modification by a user. Programs
refer to sets of instructions that tell the computer exactly how to handle a problem; they are
highly modifiable and are generally tailored to fit specific user needs.
26 FRITZ MACHLUP AND UNA MANSFIELD
Although computer science is now developing its own research identity, its
boundaries are by no means clearly defined, for it overlaps with many estab¬
lished disciplines. Zenon Pylyshyn has said that
This leads Pearson and Slamecka to ponder on what will remain in the
core of the computer-science curriculum, and they speculate that “the
knowledge-oriented component of this core will be concerned principally
with the theory of signs, sign structures, sign processes, and algorithms. The
disturbing thought is that, if this scenario is correct, the educational pro¬
grams in today’s departments of computer science . . . are not very rele¬
vant.” (Pearson and Slamecka in this volume.)
Which brings us back to the question with which we started: Whether
information and its processing, or the computer as a configuration of hard¬
ware, firmware, software, and programs, is the central object of study in
computer science.
Artificial Intelligence
External Influences
13 The term artificial intelligence was first used formally in the title of the Dartmouth Summer
Research Project on Artificial Intelligence, held during the summer of 1956. It is attributed to
John McCarthy, an AI pioneer and the principal organizer of that conference. [McCorduck,
1979, p. 96.]
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 29
artificial intelligence: They linked ideas about the workings of the nervous
system with logic and computation. To quote Minsky again,
AI grew out of the cybernetics of the 1940s with the hope that the limitations of
that methodology might be overcome by the new ideas coming from computa¬
tion—particularly the recognition that programs themselves, rather than their
applications, were objects of scientific interest. . . . The era of cybernetics was
a premature anticipation of the richness of computer science. The cybernetic
period seems to me to have been a search for simple, powerful, general princi¬
ples upon which to base a theory of intelligence. (Minsky, 1979, p. 401.)
Such questions as can computers think and are computers intelligent can be
answered only on the basis of agreed definitions of thinking and intelligence.
According to Newell and Simon, who have done ground-breaking work on
intelligent systems, “there is no ‘intelligence principle,’ just as there is no
‘vital principle’ that conveys by its very nature the essence of life.” Rather,
they point to structural requirements for intelligence, one of which is the
ability to store and manipulate symbols. “Symbols lie at the root of intelli¬
gent action, and one of the fundamental contributions to knowledge by
computer science has been to explain, at a rather basic level, what symbols
are.” (Newell and Simon, 1976a, p. 114.)
Artificial intelligence has been described as “the study of intelligence as
computation.” (Hayes, 1973, p. 40.) But there is some controversy as to
whether computation may be applied to the cognitive or subcognitive level.
[See the papers by Barr and Hofstadter in this volume.]
A distinction has been drawn between artificial intelligence and computa¬
tional psychology (simulation of cognitive processes). Margaret Boden ad¬
mits that there is “a difference in emphasis between workers who try to
make a machine do something, irrespective of how humans do it, and those
who aim to write a program that is functionally equivalent to a psychological
theory.” But she adds that
from the algorithms favored by computer scientists; and (2) its development
of list-processing program languages, when the rest of computer science was
moving toward the use of compilers.14 Allen Newell has pointed to these
isolating factors, and added a third, which became apparent as soon as the
discipline of computer science identified itself with digital computers, rel¬
egating analogue computers to a subsection of electrical engineering: (3) The
digital-computer people initially regarded computers as machines that mani¬
pulated numbers, whereas the group interested in artificial intelligence saw
computers as machines that manipulated symbols. [Newell in his lead paper
in this volume.]
According to Newell, the choice by AI workers of programming systems
as the class of systems to describe intelligent behavior, “led to psychologi¬
cally revealing tasks” and caused the AI group to turn to the discipline of
psychology for inspiration. It was a two-way partnership, with AI exerting
an influence on psychology, principally in the areas of problem-solving and
concept formation. “However, when psychology opted to concentrate on
memory structure, psychology and AI went fundamentally separate ways.
Psychologically relevant work on memory by AI researchers did exist, but
moved out of AI into psychology.” (Newell in his lead paper in this volume.)
14“List-processing languages are computer languages that facilitate the processing of data
organized in the form of lists. . . . LISP (short for LISt Processing). . . one of the most popular
of such languages . . . was developed by John McCarthy and his associates at M.I.T. during the
late 1950s and early 1960s.” (Slagle, Dixon, and Jones, 1978, pp. 778-779.) Compiling tech¬
niques for computer languages appeared with the advent of algebraic languages in the mid-
1950s. They were used originally for numerical computation; almost all of the major computer
languages are compiled (COBOL, FORTRAN, PASCAL, etc.).
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 31
Within AI itself, several subfields evolved, each with its own specific inter¬
ests, research techniques, and terminology. In addition to work on special
programming languages and tools, these included research on natural-
language understanding, vision, problem-solving, game-playing, theorem¬
proving, and so forth. It soon became clear that the main AI programs all
used the same fundamental technique—heuristic search. [Newell and Ernst,
1965.] However, “as the scope of AI programs seemed to narrow, there
arose a belief by some AI scientists that the essence of intelligence lay not in
search, but in large amounts of highly specific knowledge, or expertise. . . .
The subfield called ‘expert systems’ . . . emerged in the mid-1970s in part as
a result of this emphasis.” (Newell in his lead paper in this volume.)
As heuristic search with little knowledge of the task domain gave way to
knowledge-intensive programs, the area of knowledge engineering (com¬
puter-based management of given assemblages of “knowledge”) was born.
This has led to the development of so-called expert systems that can match
the diagnostic skills of a human expert in a relatively limited domain. How¬
ever, it is important to remember that knowledge is here being used in a very
restricted sense. To quote from The Handbook of Artificial Intelligence,
“we often talk of list-and-pointer data structures in an AI database as knowl¬
edge per se, when we really mean that they represent facts or rules when
used by a certain program to behave in a knowledgeable way.” (Barr and
Feigenbaum, 1981, p. 143.)
With the development of expert systems on computers, many workers in
AI believed that the time had come for them to move away from work on toy
tasks (small illustrative tasks, such as manipulation of blocks by a “seeing”
computer) and onto work on real tasks (e.g., construction of expert systems
for medical diagnosis). Douglas Hofstadter in his paper in this volume sug¬
gests that this move is counterproductive for the basic scientific interests of
the field.
Robotics
Cognitive Science
This is the youngest of the disciplines featured in this volume. One of its
protagonists is not sure whether this new science, or metadiscipline, is still
only “a goal rather than a reality.” (Pylyshyn in his paper in this volume.)
Very real, however, is the fact that a number of distinguished researchers
call themselves cognitive scientists, meet in conferences, publish books, and
have established a journal named Cognitive Science; its first volume ap¬
peared in 1977. Yet, this may have been a delayed christening. One of the
founding fathers took 1956 “as the year of the birth of cognitive science.”
(Simon, 1980, p. 34.) A brief description of the new discipline is in order.
Such a description may stress the subject matter, the major problems ad¬
dressed, the techniques of analysis employed, the theoretical formalisms
adopted, and a genealogical map showing from which disciplines the new
one has descended. [Collins, 1977.]
The first of these definitions refers to intelligent systems, the latter to human
mind; the idea is that both human intelligence and computer intelligence are
to be studied. (Ibid.) Alternative definitions seem to avoid the terms mind
and intelligence and instead refer to cognitive systems and mental faculties.
(Sloan Foundation, 1978, p. viii.)
The resumption of studies of the mind constitutes a decisive break with
behaviorism in psychology and neopositivism in general methodology, the
two isms that had banished all nonobservables from the domain of science.
The new area of research is the result of collaboration of members of several
disciplines—of six wider fields (represented by separate departments in
most universities) and eleven interdisciplines (or specialties) that are the
outcome of pairwise bonding between several of the major six. As shown in
the figure reproduced in the appendix to Pylyshyn’s paper, psychology,
philosophy, linguistics, anthropology, neuroscience, and computer science
are the six fields of departmental rank; computational psychology (or simula¬
tion of cognitive processes), neuropsychology, philosophy of psychology,
philosophy of language, psycholinguistics, neurolinguistics, computational
linguistics, anthropological linguistics, cognitive anthropology, evolution of
the brain, and cybernetics are the eleven subdomains of cognitive science.
Not that they are all equally active or equally productive; indeed, some of
the subdomains live only a token existence.15
“The most immediate problem areas” cultivated by cognitive science are
“representation of knowledge, language understanding, image understand¬
ing, question answering, inference, learning, problem solving, and plan¬
ning.” (Collins, 1977, p. 1.) The favored techniques of analysis “include
such things as protocol analysis, discourse analysis, and a variety of experi¬
mental techniques developed by cognitive psychologists in recent years.”
(Ibid.) Somewhat questionable is the following pronouncement about the
methodological commitments of the new discipline; “Unlike psychology or
linguistics),] which are analytic sciences[,] and artificial intelligence),] which
is a synthetic science, cognitive science strives for balance between analysis
and synthesis.” (Ibid., p. 2.) All four disciplines compared and contrasted in
this quotation are clearly empirical sciences and as such cannot do without
either analysis or synthesis. Even more puzzling is the remark by the editor
of Cognitive Science that “this discipline might have been called ‘applied
epistemology’.” Epistemology is generally regarded as a major part of
metaphysics and hence nonempirical, purely analytical; and “applied” epis¬
temology could in a way fit any branch of knowledge. Fortunately, Pylyshyn
in his paper in this volume settles the question of the methodological status
of cognitive science when he declares it to be “an empirical natural sci¬
ence.”
15 With all these subdomains and contributing disciplines, one hardly expects further additions
to the list. Yet some major researchers in the field have proposed additional shareholders in the
new enterprise: epistemology (perhaps a stand-in for philosophy) and economics [Simon, 1980,
p. 33]; and education [D. G. Bobrow, 1975.]
34 FRITZ MACHLUP AND UNA MANSFIELD
"Representation," a Keyword
for motivation and emotion appear much more analogical and continuous.”
(Simon, 1979, pp. 383-384.) Since computer simulation has been the major
research technique in cognitive science but seems unsuitable to research on
affective processes, the slow advance in the study of motivation and emo¬
tion is not surprising.
18To avoid a confusion of historical timing, we should note that, whereas some of the interdisci¬
plines involving linguistics have only recently arrived on the scene, semiotics is by no means a
young subject. “Signs,” as distinguished from “Things” and “Actions,” were among “the
three great Provinces of the intellectual World” mapped by John Locke in his famous Essay.
[John Locke, 1690.] Locke took the term semiotics from the Greek Stoics; modern semiotics
began with Charles Sanders Peirce. [Peirce, 1867.] The Swiss linguist Ferdinand de Saussure is
credited with important developments. [Saussure, 1916 and 1959.] Most widely quoted is the
exposition by Charles W. Morris. [Morris, 1939.] The studies by Roman Jakobson are cited with
special respect. [Jakobson, 1962; 1971.] A recent survey of the state of the art was published by
Thomas A. Sebeok. [Sebeok, 1976.]
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 37
Linguistics, one of the most ancient sciences, is also losing its humanistic
appearance. Mathematical linguistics has been created, which, in the manner
of Bar-Hillel, can be divided into statistical linguistics dealing with the fre¬
quency analysis of symbol systems and structural linguistics dealing with con¬
structing abstract models of language. If the first branch can be regarded as a
result of the mathematization of science, the second results from the humaniza¬
tion of mathematics: problems emerging within the humanities are formulated
in the frame of mathematics. At any rate, Chomsky’s theory of contextfree
languages is clearly a mathematical subject generated by linguistic problems.
Linguistics has also acquired some purely engineering aspects. The problems
of machine translation, working out languages for computers, and especially
the problem of “man-computer dialogue’’ have added engineering features
even to such a purely humanistic field as semantics, though the principal prob¬
lems of semantics have retained their humanistic core. (Nalimov, 1981, pp.
204-205.)
Semiotics
Realizing that the science of signs includes the science of those signs that
make up language, one cannot help seeing in semiotics the wider field that
embraces linguistics. This does not imply, however, that every semioticist
must be a linguist, nor that every linguist must be a semioticist. Division of
labor may be strongly indicated, but one expects that the two groups know
what in general the others are talking about and that both avoid using the
same terms in different meanings. The second expectation is not fulfilled, in
that one of the keywords, syntactic, is ambiguous.
In semiotics, syntactic is the level of sign theory at which the observer
finds that he or she has reason to recognize certain things as intended signs
but does not understand what they signify. In linguistics, syntactic refers to
rules or regularities in grammar regarding the coordination of words and
their functions in the structure of a sentence (like subject and predicate).
Syntactic analysis in linguistics thus represents a higher level of sign recog¬
nition than in semiotics.
Semiotics distinguishes three levels: syntactics, semantics, and pragmat¬
ics, dealing with nine kinds of signs (such as tokens, types, and tones on the
syntactic level; indexes, icons, and symbols on the semantic level). A pre¬
syntactic level (or dimension) has been proposed to cover instances where
38 FRITZ MACHLUP AND UNA MANSFIELD
the observer sees some objects, such as ink marks on paper, patterns of
chiseled grooves on rocks or stones, or tree stumps in the woods, but does
not know whether they are signs; that is, artifacts that signify something, at
least to the person or people who produced them and wanted to communi¬
cate to others. [Seiffert, 1968, pp. 80-81.] One speaks of the syntactic di¬
mension when one clearly recognizes that the observed objects are signs but
does not know what they signify. When what is spoken, written, or printed is
in a foreign language that the observer does not understand, the signs are
still in the syntactic dimension. The borderline of the semantic dimension is
blurred when communication is in the recipient’s own language but on a
technical level not comprehensible to him, like an argument in the jargon of
another scientific discipline. The reader or listener may understand almost
every word, and even its grammatical function in each sentence, but still not
grasp the meaning of the sentences. To be at the truly semantic level, those
at the receiving end have to be somewhat familiar with the subject matter
and/or with the linguistic peculiarities in which it is treated. For example,
native Germans with twelve or fourteen years of schooling may fail to under¬
stand Kant’s Critique of Pure Reason if they have had no background in
philosophy. While these readers are on the semantic level with regard to the
words of the text, they are only on the syntactic level as far as Kant’s total
message is concerned. [Seiffert, 1968, pp. 84-85.] In the pragmatic dimen¬
sion, semantic comprehension is associated with action, as when the mes¬
sage is an invoice requiring payment or a summons to appear in court.
What contributions has semiotic research made to either the study of
natural language or to studies of information? This question is debated.
Views expressed in various papers in this volume show disagreement. Pear¬
son and Slamecka hold that semiotics has promoted progress in informatics.
According to a skeptical view, “the accomplishments of the research
program in general semiotics . . . have not been spectacular, unless one
classifies the achievements of cybernetics under this heading.” (Weinreich,
1968, p. 169.) As to linguistics, the same source has this to say: “Although
no sign system equals language in the variety and overlapping of semiotic
devices employed, it has been instructive to embed the study of natural
language in a broader investigation of sign phenomena of all kinds, including
substitutes for language (e.g., flag codes) and extension of language (gesture
patterns, chemical formalism, etc.).” (Ibid.)
The old controversy about heredity versus environment (nature versus nur¬
ture) as major determinant of human intelligence reappeared in the field of
linguistics: Was the understanding of, or feeling for, grammar entirely a
result of learning or was there a predisposition for it due to genetic evolu¬
tion? The terms chosen for these notions are strangely different from those
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 39
Cybernetics
More than any other discipline, cybernetics sprang from the head of one
founding father, Norbert Wiener.19 [Wiener, 1948 and 1961.] He regarded
cybernetics as “a new scientific subject,” which at the time of the second
edition was “an existing science.” (Ibid., p. vii.) The question is only
whether the scope of that science coincides with that of Wiener's book, in
the sense that cybernetics comprises all that is in Cybernetics, but not more.
Wiener himself held that “the center of interest” in a subject that has “real
vitality” is apt to “shift in the course of years.” (Ibid.) Hence, cybernetics
should not be limited to what Wiener put down in his book. On the other
hand, not all that is covered in the book can reasonably be said to be
cybernetics, since he devoted many pages and several chapters to applica¬
tions of cybernetics to various empirical subjects, such as computing ma¬
chines, psychology, psychopathology, social organization, learning theory,
electrophysiology, and what not.
A recent leaflet by the American Society for Cybernetics contains a mot¬
ley collection of definitions of cybernetics. After giving Wiener’s formula¬
tion (from the title of his book) and Ampere’s obsolete suggestion of the
science of government, the leaflet supplies four definitions ranging from
epistemology via organization theory and the science of form and pattern to
the art of manipulating defensible metaphors."0
19It is easy to show that some of the fundamental ideas presented in Wiener’s Cybernetics—
say, the notion of (though not the term) feedback—had been known to earlier writers about
reflexes and related physiological phenomena. Also, Bertalanffy’s general system theory, pre¬
sented in German articles, anticipated Wiener’s publication. Still, the exposition of these ideas
under the name cybernetics is unquestionably Wiener’s.
20To quote in full, “For philosopher Warren McCulloch, cybernetics was an experimental
epistemology concerned with the communication within an observer and between the observer
and his environment. Stafford Beer, a management consultant, defined cybernetics as the
science of effective organization. Anthropologist Gregory Bateson noted that whereas previous
sciences dealt with matter and energy, the new science of cybernetics focuses on form and
pattern. For educational theorist Gordon Pask, cybernetics is the art of manipulating defensible
metaphors, showing how they may be constructed and what can be inferred as a result of their
existence.” (From a leaflet published by The American Society for Cybernetics, 2131 G Street
NW, Washington, D.C., 20052.)
40 FRITZ MACHLUP AND UNA MANSFIELD
Murray Eden in his paper in the present volume considers two related
insights as the most important findings of cybernetics: First, that there is an
essential unity in the set of problems in communication, control, and statisti¬
cal mechanics (noisy phenomena), whether they are to be studied in the
machine or in living tissue. Second, that “the computing machine must
represent almost an ideal model of the problems arising in the nervous
system.” (Eden in this volume.)
Connections or Applications?
Eden has been able to show connections, including some close ones, be¬
tween cybernetics, either as a science or (more often) as a viewpoint, and
some thirty other disciplines or specialties. This many-sidedness alone
should suggest that the connections are not empirical conjunctures of phe¬
nomena that are common subjects of observation and explanation by the
different disciplines but, instead, that cybernetics offers a technique of
analysis applicable to a large variety of phenomena, natural, social, and
man-made. Without hinting at the possibility of cybernetics being a
methodological research program of wide applicability, Eden comes to this
conclusion: “The notions of cybernetics have permeated many disciplines—
computer science, information theory, control theory, pattern recognition,
neurophysiology, psychophysics, perceptual psychology, robotics, and the
like. Having been integrated into them, cybernetics has performed the func¬
tion for which it was proposed.” (Eden in this volume.)
This conclusion sounds as if cybernetics has fulfilled its mission and can
be buried—which would not be the appropriate thing to do in the case of a
successful and fertile, methodologically sound scientific research program.
Perhaps another interpretation offered by Eden as well as by Elias, Arbib,
and Mattessich explains the virtual disappearance of cybernetics from the
Western academic establishment. According to that interpretation, cyber¬
netics is a part of general system theory and has been completely absorbed
by this much broader scientific research program.
When Arbib holds that “the history of cybernetics is, in large part, also
the history of cognitive science,” one might infer that he sees cybernetics as
an empirical discipline like cognitive science. The same impression is con¬
veyed when he reports that members of his research group used “cybernet¬
ics to refer to a conjoined study of brain theory and AI”—both clearly
subjects of empirical research. On the other hand, Arbib speaks of modern
system theory as “the descendant of Wiener’s control theory”—and this
agrees with our conception.
gent properties; structure, hierarchy, and evolution; function and goal orien¬
tation; information, control, and feedback; environment and its influence;
system laws and mathematical homologies. According to Mattessich, cyber¬
netics is most appropriately regarded “as a subset of systems research.’’-1
We endorse this view.
The tendency to link cybernetics with general system theory and, even
further, with information theory began in Europe almost as soon as Norbert
Wiener’s and Claude Shannon’s works were reported. Thus, in a series of
conferences held in Paris in the spring of 1950 on Cybernetique: theories du
signal et de l’information, the first speaker defined cybernetics as “the sci¬
ence of relations, controls, and transmissions of information.’’ (de Broglie,
1951, p. v.) This definition announces the unification of the triad: The
science of relations is general system theory, the science of controls is
cybernetics proper, and the science of transmissions of information is infor¬
mation theory. The speaker subsumed them all under the name cybernetics.
This merger of the three disciplines under a single name can be seen on
the organization charts for the academic enterprise in Europe, but not in the
United States. We doubt that most American representatives of general
system theory and information theory would be prepared to accept such a
take-over offer. On the other hand, a take-over of cybernetics by system
theory seems to be more or less a fait accompli.
System Theory
21 Some writers go farther and make the two disciplines one and the same: “Cybernetics is the
science of general system theory.” (Paul E. Martin, 1981, p. 11.) This, however, is the view of a
writer based in Central Europe, who thinks that cybernetics serves chiefly to devise systems of
“improved, simplified and less costly information handling [by management] in business and
government,” and that Stafford Beer, who wanted to design the socialist planning system of
President Allende in Chile, is “one of the greatest living cyberneticians of our time.” (Ibid., p.
14.) As Eden and Elias mention in their papers in this volume, the meaning of cybernetics in the
Soviet Union is wider still in that it comprises mathematical economics and econometrics.
42 FRITZ MACHLUP AND UNA MANSFIELD
esting but not convincing. For we established that the same writers alter¬
nated in casual inconsistency between using plural and singular interchange¬
ably in the same papers, chapters, pages, and even paragraphs. (Why were
they not queried by their editors? We, too, decided to grant our contributors
freedom to be inconsistent in this matter.)22
The question which noun is modified by general in general system theory
proved even more bewildering. Was it the general theory of systems or the
theory of general systems? At one point, we became convinced that the
absence of a hyphen between the first two words was not deliberate but mere
sloppiness. Our evidence seemed strong: The yearbooks of the Society for
General Systems Research have carried the title General Systems, and au¬
thors wrote of general systems as contrasted with special systems. We were
disabused of this conclusion when we saw foreign-language renditions of
general system theory—all emphasizing the generality of the theory, not of
the systems.23 We should not be surprised, however, if a parishioner of one
of the various denominations of system theory were to castigate us for
downgrading the theory of general systems. The ferocity with which differ¬
ent system theorists attack one another’s methods and methodological posi¬
tions has shown us that whatever is said about systems will make some of
the partisans very angry.
There is no agreement about definitions of either system or system
theory.24 This need not bother us: Scholars (scientists) may disagree on how
to define a term and yet agree on its meaning. They are too proud to imitate
the layman’s excuse, “You know what I mean,” when he is aware of failing
to express himself clearly. And scholars are usually able to convey clear
meanings without formulating explicit definitions; not so when they speak
about systems. Still, three or four notions are mentioned as fundamental by
"One of the founding fathers of this subject matter, Ludwig von Bertalanffy, in his book on
General System (singular) Theory wrote of the systems (plural) approach, and he reported that
the proposed name of the Society for General System (singular) Theory was changed to Society
for General Systems (plural) Research. [Bertalanffy, 1968, pp. 4, 15 ] In the foreword to his
book, General System (singular) Theory, he used Systems (plural) Theory in the second sen¬
tence and in the next sentence, stated that “Systems (plural) Science, or one of its many
synonyms, is rapidly becoming part of the established university curriculum.” (p. vii.) We can
easily confirm this last statement: The system theorist George Klir professes his discipline in a
Department of Systems (plural) Science. In support of those who favor the singular, we may
mention that the Theory of Games (plural) is always rendered as Game (singular) Theory;
likewise, the Theory of Numbers (plural) is transformed into Number (singular) Theory. On the
other hand, one speaks of a numbers (plural) game, and economists talk about commodity
markets and commodities markets without awareness of a possible distinction.
23Five years before his article “An Outline of General System Theory” [Bertalanffy, 1950],
Bertalanffy published in German a paper “Zu einer allgemeinen Systemlehre.” [Bertalanffy,
1945.] Similarly, we find “La theorie generale des systemes” in French and “La teoria generale
dei sistemi” in Italian.
24 “Definitions of general systems theory differ. . . . The divergence begins at the very start
when attempts are made to define a system.” (Rapoport, 1976, p. 11.)
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 43
most system theorists: a whole (or assemblage, set, group, etc.), parts (or
elements, objects, constituent entities, members, etc.), relations (among the
parts), and sometimes, as an afterthought, environment (outside the bound¬
aries of the system or also within it for each entity).
The biggest question arises with regard to the relations. Some writers
specify interactions and thereby restrict the class of possibly relevant rela¬
tions; others say interdependence, which restricts it less, but some speak of
interrelations, which includes much more than interdependence and interac¬
tions. Relations among the entities may be of many kinds—for example,
logical (contrariety, transitivity, etc.), spatial (next to, below, etc.), com¬
parative (larger, faster, etc.), attitudinal (friendly, uncooperative, etc.), hier¬
archical (foreman, boss, etc.), kinship (son, sister, etc.)—without necessar¬
ily implying particular influences, stimuli, responses, actions, or reactions.
Most of the concrete systems and special systems that are presented in the
literature (as illustrations of the nature and significance of general systems)
feature mutual influences, attractions, communications, and interactions.
Exceptions, however, are not infrequent, and when the argument is pre¬
sented in mathematical language, with any one element changing as a func¬
tion of changes of all others, one can no longer make out the nature of the
interrelated parts and the nature of their interdependence. There may be
humans (acting on command, request, advice, or in spite; or in order to gain
or to survive); cells (in a living body); neurons (firing at neighboring ones);
species of animals (multiplying or dying out or evolving physiological, ana¬
tomical, or behavioral traits); molecules in different arrangements (say, in
chemical substances); electrical impulses (transmitted over a channel). But
the parts of the system may also be shapes and colors (on a canvas or
screen); tones in various intervals (in melodic or harmonic consonances and
dissonances); ink spots on pieces of paper; all sorts of patterns and
configurations, arrangements, and organizations. Indeed, it is the very idea
of a general theory of systems in general to say something about all kinds of
entities that are interrelated members of a set (which may be a subset of
another set) or of a whole (which may be part of another whole).
The question is whether much of great importance can be said about all
possible systems.25 Important in this context may mean helpful to inves¬
tigators in several disciplines in their research, analysis, and interpretation;
or it may mean helpful to organizers and managers in the practical world ot
engineering, business, and government; or helpful to philosophically minded
scholars in their abstract reflections about the scientific enterprise in
general.
25 “.. . all we can say about practically everything is almost nothing.” (Boulding, 1956b, p. 197.)
44 FRITZ MACHLUP AND UNA MANSFIELD
Most writers on system theory see the value of their formal theory in its
aid to various other disciplines, formal or empirical. Particularly, general
mathematical systems theory is said to provide foundations for other disci¬
plines, to “offer both a conceptual framework and a working methodology.”
(Mesarovic, 1968b, p. 60.) According to one authority, the logical connec¬
tions among phenomena, model, and system are expressed in these two
statements: “(a) A theory of any real-life phenomena ... is always based on
an image, termed a model, (b) Without introducing any constraints what¬
soever the formal, invariant, aspects of that model can be represented as a
mathematical relation. This relation will be termed a system.” (Ibid.)
26 Waterman went on to say that “[djepending on the available data and the purpose of the
analysis any, or more usually several, of a wide range of specific techniques may be employed
ranging from the use of information theory and cybernetics to computer simulation or multivari¬
able statistical analysis.” (T. H. Waterman, 1968, p. 4.)
46 FRITZ MACHLUP AND UNA MANSFIELD
Information Theory
Those who have read ten or more different expositions of this field of re¬
search (and have not been influenced by a charismatic teacher) cannot help
being uncertain about the field’s real subject, it scope, and its name. Infor¬
mation theory is only one of several alternative names of this discipline;
among other designations are mathematical theory of communication, com-
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 47
One reason for the diversity of claims is that “ ‘Information theory’ is used
in at least three senses. In the narrowest of these senses, it denotes a class of
problems concerning the generation, storage, transmission and processing of
information [in a very special sense], in which a particular measure of infor¬
mation is used.” (Elias, 1959 and 1968, p. 253. Emphasis added.) This is
essentially Shannon’s theory.28 [Shannon, 1948.] “In a broader sense, infor-
27Some of these positions may be exemplified here by direct quotation: "Unlike Newton’s laws
of motion and Maxwell’s equations, which are strongly physical in that they deal with certain
classes of physical phenomena, communication theory is abstract in that it applies to many sorts
of communication, written, acoustical, or electrical. . . . Communication theory proceeds from
clear and definite assumptions to theorems concerning information sources and communication
channels. In this it is essentially mathematical, and in order to understand it we must under¬
stand the idea of a theorem as a statement . . . which must be shown to be the necessary
consequence of a set of initial assumptions. This is an idea which is the very heart of mathemat¬
ics as mathematicians understand it.” (Pierce, 1961 and 1965, p. 18.)
On the other hand, one of the contributors to this volume contends that “Shannon’s theorem
about the existence of codes that make it possible to transmit information at rates that come
arbitrarily close to a specified upper limit with arbitrarily low probabilities of undetected error
rooted what others called ‘information theory’ firmly as an engineering discipline.” (Kochen in
his paper in Section 5.) Strangely enough, the same author a few years earlier had characterized
information theory (of the Shannon type) as a mathematical specialty.
In contradistinction to this view, a Russian mathematician stated that “Information Theory
is one of the youngest branches of applied probability theory.” (Khinchin, 1956; English trans¬
lation, 1957, p. 30.)
A less apodictic position is taken by another of the contributors to this volume. He holds
that after “the popularity of the information-theoretical approach, especially in the 1950s, [the]
later disillusionment with that approach, especially among psychologists, stems from an initial
lack of public clarity as to the scope and limits of Shannon’s theory . . . , and especially from a
widespread failure to distinguish the concept of information per se from various measures of its
‘amount’. It should be added that some of the most generally useful qualitative concepts of
information engineering . . . , such as feedback and feedforward, have quite independent origins
and owe nothing to the mathematical theory of information measurement.” (MacKay in his
paper in this volume.)
28 “This area is also called ‘coding theory’ and . . . ‘the mathematical theory of communica¬
tion’.” (Elias, 1968, p. 253.)
48 FRITZ MACHLUP AND UNA MANSFIELD
29That von Neumann continued to have misgivings about speaking of information and informa¬
tion theory may be inferred from the fact that he placed quotation marks around information
theory” when he first referred to it in the introduction of his posthumously published booklet
The Computer and the Brain, [von Neumann, 1958, p. 1.]
30“Communicating, from an engineering point of view, means simply moving electronic traffic
50 FRITZ MACHLUP AND UNA MANSFIELD
results a theory of communication and he should not have allowed his fol¬
lowers to call it information theory.
Regret about the misleading name given to the theory and its keyword is
shared by many eminent writers on the subject. It is a pity, said Colin
Cherry, “that the mathematical concepts stemming from Hartley have been
called ‘information’ at all.” (Cherry, 1957 and 1966, p. 50.) What is really
involved in the mathematical theory of communication are signs and signals
and their transmission over channels of communication.
from one place to another. It matters little if the signal represents random noise or a Shake¬
spearean sonnet.” (Branscomb, 1979, p. 143.) We wish to record dissent from this statement by
a professor of communication engineering: He emphasized that his discipline is very much
concerned with the distinction between signals and noise.
The first signal reduces the possible choices from 8 to 4; the second signal, from 4 to 2; and the
third signal commands the final selection. That three binary (or yes/no) signals suffice to select
one message from an alphabet of eight is mathematically expressed by the fact that 8 = 23.
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 51
systems into discrete, equally spaced ordinates, and only samples of succes¬
sive amplitudes are transmitted to the receiver. The composite signal re¬
ceived is, within limits, the same as the unsampled signal supposedly sent.
[Ibid., p. 142.] (Do we risk ridicule from experts if we suggest as an analogy
the sequence of still photographs in film strips showing apparently continu¬
ous movement?)
According to some specialists, the greatest contribution to statistical com¬
munication theory was Shannon’s capacity theorem, also called the coding
theorem, regarding noisy channels. It states that within certain limits, de¬
pending on specified channel constraints, it is possible “by using codes of
sufficient length to set the input rates arbitrarily close to [the channel’s
capacity] and yet maintain [despite the noisy channel] the probability of
reception error as small as desired.” (Feinstein, 1958, p. 43.) Although elec¬
trical engineers had known this before Shannon’s equations, they agree that
the formal and rigorous determination of capacity limits is of fundamental
importance.
In these few paragraphs on elements of communication theory—signal
transmission and channel capacity—we have avoided using the word infor¬
mation, but it is time to explain what the engineers mean by it. They do not
mean the what that is to be communicated, but, instead, they mean the
instruction that the sender, by signals, conveys to the receiver, commanding
it (not him or her, but it) to select a particular message from the given
ensemble of possible messages. Thus, this information in the engineering
sense is an instruction to select.32 And the rate of information is the rate at
which such instructions can be transmitted per signal or per second.
[Cherry, 1957 and 1966, pp. 170-180.] However, even the word instruction
is misleading, in that it may produce an association with verbal instructions
given to a person equipped with a mind and supposed to engage in a cogni¬
tive process. The word command has the same semantic defect. Activation
would be a more appropriate term: Activating impulses are transmitted
through the channel to the receiver. Thus, activation rate might be a suitable
substitute for information rate.
One may wonder why people other than communication engineers were
willing to accept the engineering sense of information for contexts where no
bandwidth, no signals per second, no channel capacity, no signal/noise ratio,
32“Mathematical communication theory concerns the signals alone . . . abstracted from all
human uses” and apart from all meaning. The signals control selection of signs from the
receiver’s alphabet, as when “one teletype machine [is] communicating with another. At the
transmitting end, the operator selects and presses keys one at a time; coded electrical signals
are thereby sent to the receiving machine, causing it to select and depress the correct key
automatically. We see the receiver keys going down, as though pressed by invisible fingers.”
(Cherry, 1957 and 1966, pp. 168-169.)
52 FRITZ MACHLUP AND UNA MANSFIELD
and so forth, were involved. When scholars were chiefly interested in cogni¬
tive information, why did they accept a supposedly scientific definition of
“information apart from meaning”? One possible explanation is the fact that
they were impressed by a definition that provided for measurement. To be
sure, measurement was needed for the engineering purposes at hand; but
how could anybody believe that Shannon’s formula would also measure
information in the sense of what one person tells another by word of mouth,
in writing, or in print?
We suspect that the failure to find, and perhaps impossibility of finding,
any ways of measuring information in this ordinary sense has induced many
to accept measurable signal transmission, channel capacity, or selection
rate, misnamed amount of information, as a substitute or proxy for informa¬
tion. The impressive slogan, coined by Lord Kelvin, that “science is mea¬
surement” has persuaded many researchers who were anxious to qualify as
scientists to start measuring things that cannot be measured. As if under a
compulsion, they looked for an operational definition of some aspect of
communication or information that stipulated quantifiable operations. Shan¬
non’s formula did exactly that; here was something related to information
that was objectively measurable. Many users of the definition were smart
enough to realize that the proposed measure—perfectly suited for electrical
engineering and telecommunication—did not really fit their purposes; but
the compulsion to measure was stronger than their courage to admit that
they were not operating sensibly.33
A British scientist had also been searching for definitions that permitted
measurement of information: Donald MacKay. He came up with three dif¬
ferent definitions of measurable information, some of which had been pro¬
posed by statisticians and engineers. [Fisher, 1935; Gabor, 1946.] He distin¬
guished the logon content, the metron content, and the selective-information
content, the latter being Shannon’s measure of information. In his paper in
this volume, he writes: “It would be clearly absurd to regard these various
measures of ‘amount-of-information’ as rivals. They are no more rivals than
are length, area, and volume as measures of ‘size.’ By the same token it
would be manifestly inept to take any of them as definitions of information
itself.” With regard to the Shannon definition (in terms of probability or, as
MacKay put it, unexpectedness), he has this to say: “To try to translate
every reference to information (whether in biology or elsewhere) into a
statement about unexpectedness would be as inept, and as conceptually
Procrustean, as translating all references to a house into statements about
The irrational zeal to measure even where it makes no sense is frankly expressed in the
following statement: “In many ways it is less useful to measure the amount of information [in
Shannon’s sense] than the amount of meaning. In later chapters, however, 1 reluctantly deal
more with measurement of the amount of information than of meaning because as yet meaning
cannot be precisely measured.” (J. G. Miller, 1978, p. 12.)
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 53
its size.” MacKay’s exposition, however, was not without linguistic in¬
felicities. He was very helpful when he warned—as early as 1950—against
confusing information in the original sense of the word (where some mean¬
ingful content is expected) with amount-of-information (the latter referring
to measurable improbability or unexpectedness of signals from a given set).
But then he introduced or accepted the term information-content—which a
commonsensible reader would expect to refer to the meaning-content of the
information—and wanted that term to denote a measurable amount of infor¬
mation-transmission, regardless of any meaning-content.34 [MacKay, 1951a;
1954b; 1969, pp. 56, 156-160, 163-167.]
Confusing Explanations
34 Not only general readers but also logicians mean by contents of information their semantic
contents. Thus, Bar-Hillel and Carnap held that the terminology and the theorems of telecom¬
munication research were unfortunately applied “to fields in which the term information was
used ... in a semantic sense, that is, one involving contents or designata of symbols....” (Bar-
Hillel and Carnap, 1953b, p. 147. Emphasis added.)
54 FRITZ MACHLUP AND UNA MANSFIELD
less or more) questions and the transmitters select the answers, though
before the questioning begins, they have also selected the correct answer to
be guessed. The relevance of all this to the process of signal transmission is
not made clear in any expositions that we have read.
The core of the theory is probability, yet few expositors care to state just
what is probable about a message transmitted. A communication system
such as the telegraph or the telephone does not usually transmit messages
that tell which side of the coin happened to come up or which number of
eyes (or dots or spots) were on the top of the die or which among the playing
cards in the deck was picked. Messages conveyed by telegraph are usually in
the form of strings of words (in a natural or artificial language) and numbers.
It is easy to understand that the signals adopted to represent letters in a
telegram should be chosen so that simpler signals stand for the most fre¬
quently occurring letters. A coding convention of this sort will reasonably be
based on the probability of particular letters being required to compose
words—where probability means the frequency of these letters in the
English (French, German, etc.) language. (The most frequent are e and t,
then i, a, n, and m.)
The frequency of individual letters occurring in an English text (or an¬
other language) matters a good deal when engineers designing a telegraph
system determine what signals ought to be used in encoding, transmitting,
and decoding. The most frequently used letters should be represented by the
shortest sequence of signals, with longer sequences of signals reserved for
less frequently used letters. In most expositions of the theory, discussion of
this encoding problem is followed by a discussion of doublets (digrams) and
triplets (trigrams), that is, sequences of two or three letters that occur fre¬
quently, infrequently, or never in the particular natural language. The bear¬
ing of these different frequencies, however, is less on the assignment of
signals to signs, but rather on the ease of detecting errors in scrambled
telegrams. (For example, sck is not likely to appear in any English word.)
Such detection, however, is unlikely to be built into the architecture of the
system but rather into the cognitive capability of the human recipient.
Turning from telegraph to telephone communication, the physical signals
used to transmit words spoken by a human voice into a microphone and
channeled through wires and cables are surely very different, and the proba¬
bility of particular letters in a written language is no longer significant. Now,
differences in pitch may be important. (We have asked ourselves whether a
telephone system optimal for Chinese speakers would be the same as one for
English.) We can send music over the telephone, and the relative frequency
(probability) of very high, very low, very loud, and very soft tones probably
matters. These issues, however, are not ordinarily included in elementary
expositions of information theory and the so-called amount-of-information
transmitted. The illustrations furnished, even by those who repudiate any
concern with semantic content, usually employ sentences that do have
CULTURAL DIVERSITY IN STUDIES OF INFORMATION 55
meaning, and this further confuses the reader. For example, when it is
pointed out that the probability of a message depends on how many possible
messages there are, and when this is followed by an example of two English
sentences (widely used in practice manuals for students of typing), the
reader will be puzzled about the information contained in either of the two
messages that was selected. Was this a selection by chance, and is each of
the messages equally probable in any other sense? [Elias, 1959 and 1968, p.
254.]
When we are told about the probability of a message, we remain puzzled
about just what is probable about a message. We cannot help (despite all
warnings) thinking first of the probability of its proving true, useful, or
valuable. Second, we think of the probability of its being selected by a
transmitter making a rational choice, either trying to be helpful, perhaps
obeying a command, or trying to deceive the recipient. Third, remembering
information theorists’ advice to forget about meaning and not to forget about
the given pool of possible messages, we think of the probability of the
message being picked at random from all possible messages. Fourth, having
learned about disturbing noise being mixed with genuine signals, we think of
the probability of the message reaching the receiver without undue noise
and, thus, as it is intended by the sender. Finally, backsliding into the
normal attitude of connecting messages with something they are supposed to
convey, we think of the probability of the message being understood by the
recipient. This half-dozen meanings of the probability of a message does not
exhaust the list, and the trouble is that formulations chosen by most ex¬
positors do little to help the recipient of their messages grasp the entire
exercise.35
With the difficulties of properly interpreting the many inept and obscure
expositions of information theory, added to the fact that there is not one
information theory but several of different scope and even different objects,
it would be an unduly daring attempt to characterize the methodological
nature of the insufficiently specified discipline. The question, for example,
whether it is an empirical or an entirely analytical theory cannot be an¬
swered before one decides just what the theory says and what it denies.
35 In one of Cherry’s formulations, we are confronted with three probabilities: “the probability
of message x being sent, assessed from past observations of the transmitter ; the probability
of an x being sent, on those occasions when y is received”; and “the likelihood that, if any
particular x had been sent, the specific y would be received. ’ (Cherry, 1957 and 1966, p. 202.)
We interpret this argument as an exercise on deriving the likelihood—that is, the so-called
probability of a single event of a specific type—from prior statistical probabilities, as frequen¬
cies observed in the past, where the first is the frequency of x alone and the second, the
frequency of a conjuncture of x and y.
56 FRITZ MACHLUP AND UNA MANSFIELD
[Editor's note (U.M.): As well as making suggestions for improving various parts of the
Prologue, George Miller provided us with a comment on its section dealing with cognitive
science. This was contained in a letter to Fritz Machlup dated January 18, 1983, and Professor
Machlup’s immediate reaction was: “We should publish it as it is.” However, he wanted more
time to think about it, but he died unexpectedly before we had had a chance to discuss it again. I
have decided to be guided by his initial reaction, and Miller has kindly agreed to have the letter
published as it is, with only very slight editing.]
Dear Fritz:
It has taken longer than I had hoped to read “Cultural Diversity in Studies of
Information” by you and Una. I hope my impressions will still be of some
interest to you.
I enjoyed most of it, although at points I must confess that your patience with
the territorial claims of the different cultures is greater than mine. I particularly
relished your defense of pluralism, and your interest in good ideas regardless of
whether they come labeled as “science” or “practice.” As one who lives in a
marginal science myself, I long ago discovered that attempts to work solely in
the manner that I had been taught to regard as characteristic of true science
could only stunt my intellectual growth as a psychologist. So I decided to
pursue problems wherever they led; if the result was not science, that was
unfortunate but not important ....
Your discussion of cognitive science seems to me to rely too heavily on the
views of those who take computers and artificial intelligence as providing the
definitive characteristics of this field. Personally, I find it difficult to communi¬
cate clearly my enormous debt to people like Simon, Newell, and Minsky for
helping cognitive psychology take a great leap forward, yet at the same time to
communicate with equal clarity that the human brain/mind system is enor¬
mously more complicated than, and different from, any contemporary compu¬
tational system.
An interesting point that you make might help me straighten out my difficulties
with computer-oriented cognitive science. In my opinion, computer science
applied to the problems of cognitive psychology should yield cognitive psy¬
chology. The fact that it has not, that it has yielded something novel—concepts
of mind and intelligence so abstract that it doesn’t matter whether the system
possessing them is living or dead—may mean that I have mistakenly regarded
computer science as a formal science.
But I suspect that the tensions already tearing the new field apart are worse
than that diagnosis might suggest. Contemporary computers manipulate sym¬
bols, and symbol manipulation is certainly one kind of intelligence. But
whether that kind of intelligence is the only kind remains an open question. I
can accept the claim that the brain must be a universal Turing machine, but that
58 FRITZ MACHLUP AND UNA MANSFIELD
tells me almost nothing as long as I don’t know what kind of universal Turing
machine it really is. (It is like telling me that the correct theory of the brain’s
function could be written in ink.) Unfortunately, I am convinced the brain is
not the same kind of universal Turing machine as are the machines presently
merchandised by IBM.
At this point, therefore, I have personally taken refuge in the plural s that you
endorse so strongly .... The cognitive sciences are those scientific disciplines
sharing an interest in the representation and transformation of knowledge (read
“information” in the present context). But, as you have pointed out some¬
where, sharing an interest and being a single discipline are very different
things. Moreover, my definition of the cognitive sciences is not easily distin¬
guished from your definition of the information sciences, yet it is plain enough
that your information science is a larger, more catholic assemblage than my
cognitive science. At which point I merely shrug. You, I believe, feel a com¬
mitment to rationalizing the difference.
of precategorical memory for the “raw feels" of life presumably requires some
kind of analogue representation, but whatever it is, contemporary computer
science does not help us understand it any better. In other words, the impor¬
tant split here is not cognition vs. conation and emotion, but rather a split
within cognition itself: symbolically represented cognition vs. primal, nonsym-
bolically represented cognition. In short, computer-oriented theories of cogni¬
tion are not so much wrong as incomplete ....
Thank you, both of you, for letting me see this. I hope these random associa¬
tions will be of some help.
Sincerely,
George A. Miller
.
,
SECTION 1
INFORMATION SCIENCE
Its Roots and Relations as Viewed
from the Perspective of Cognitive Science
Zenon W. Pylyshyn
were very much part of the same intellectual movement. These technical
achievements included the following closely related developments:
some, like me, feel that it is much more than a metaphor) than those that had
been prevalent in earlier attempts to unite the psychological, biological, and
physical modes of explanation. The historical and philosophical issues sur¬
rounding this discovery are beyond the scope of this essay; however, they
are discussed elsewhere in some detail. [Pylyshyn, 1970; Arden, 1980u; and
Pylyshyn, 1982.]
The overlap among these four areas, as measured in terms of the propor¬
tion of people who contributed to them, is remarkable. For example, the
earliest writings on the theory of interdependent competitive decisions
(theory of games), while dating back to the 1920s, became a central concern
in the 1940s, about the time when their author, John von Neumann, became
involved in designing the earliest stored-program electronic digital computer
[Pylyshyn, 1970] and in applying notions of computation to problems of
biology [Burks, 1970]. Claude Shannon also wrote some early papers on
computing and the use of symbolic logic in the design of switching circuits.
Norbert Wiener, who had been a student of Bertrand Russell, contributed to
technical developments in both information theory and the theory of control
systems. Even more important was Wiener spearheading the rapidly grow¬
ing view that, at a suitable level of formal abstraction, these ideas were also
relevant to understanding any complex system, including social and biolog¬
ical systems. This led to the view that there are important abstract principles
that are shared by any highly interacting sets of functional elements, regard¬
less of whether these are sets of cells in a liver or brain, sets of animals in
an ecological system or sets of humans in a social system or sets of elec¬
tronic components wired together to form a complex artifact. This view gave
rise to an area of study sometimes referred to as general systems theory.
Analog computation was also closely tied to both control theory and
information theory. The mathematical tools for analyzing analog systems—
calculus and differential equations—were the ones that were relevant to
continuously varying electronic systems such as those that occur in feed¬
back control loops. By contrast, the study of digital computers, decisions,
and the parts of information theory having to do with encoding were often
discrete. As it turned out, discrete systems were found to be mathematically
more tractable and also much more interesting in a number of respects. For
example, they appear to be far more general (inasmuch as they contain the
class of universal computers, or Turing machines) than continuous ones.
The common mathematical tool that was used in these cases was symbolic
logic.
Almost all early workers on computers and communication systems were
trained in symbolic logic. Indeed, even to this day, mathematical logic has
had a central role in developing ways of conceptualizing problems in the
entire spectrum of information-related disciplines. Despite some early criti¬
cisms of the predominance of logic in the study of artificial intelligence, logic
is becoming more central than ever—both for the analysis of programming
systems and as a formalism for representing knowledge and even programs
themselves.
INFORMATION SCIENCE VIEWED FROM COGNITIVE SCIENCE 67
The one area that has not been mentioned already but which played an
important, though initially highly speculative part in these developments,
concerns the relation of this network of ideas to understanding human
thought and the nature of intelligence. As early as 1950, attempts to apply
certain ideas developed in the study of foundations of mathematics (for
example, the notion of computational mechanism as exemplified by the
Turing machine) to understanding the nature of intelligence were published
by Turing, von Neumann, Shannon, and others. But it was with the avail¬
ability of actual computing devices in the early 1950s that this goal began to
be pursued in earnest.
There were two primary strands in this early work. One was directed at
the class of self-organizing or adaptive systems and emphasized learning,
statistical-pattern recognition, and the modeling of neural networks. This
strand frequently associated itself with the term cybernetics. The second
strand was more molar. It was directed at understanding problem-solving
and thought processes at what we would now call the cognitive or the knowl¬
edge level. These two strands were nearly equally represented at a historic
symposium held in 1958 at the National Physical Laboratory in England on
Mechanization of Thought Processes. In the last fifteen or so years, how¬
ever, work on the class of approaches that are concerned with self¬
organizing systems—especially those influenced by speculation about the
properties of randomly connected neural networks—has all but disap¬
peared. On the other hand, work on the second class of approaches has
grown rapidly and also changed in emphasis in order to reflect a new under¬
standing of the problem of intelligence. The first decade of this work
(roughly 1955-1965) was dominated by the search for general problem¬
solving techniques together with powerful methods called heuristics for con¬
trolling the breadth of trial-and-error search. This work is well represented
by the reports reproduced in a classical book of readings. [Feigenbaum and
Feldman, 1963.]
The second decade [described in Minsky, 1968; Feigenbaum, 1969] was,
by contrast, more concerned with problems of planning and designing good
forms of knowledge representation and the development of high-
performance (or expert) systems in narrower domains. The key terms
changed from learning, classification, and search to planning, knowledge
representation, and semantics. The most recent general picture of the state
of the field generally referred to as artificial intelligence can be found in
a chapter in the collective volume What Can Be Automated? [Arden, 1980a.]
gence. It is not the only form in which the search for an understanding of
mind is proceeding. Nor, furthermore, is this effort being carried out in only
one academic discipline, for example, the one calling itself cognitive psy¬
chology or artificial intelligence. There are many distinct disciplines con¬
tributing to the effort. What characterizes this particular class of ap¬
proaches—which have recently come to be collectively (and somewhat
loosely) called cognitive science—is an allegiance to the network of ideas
that I roughly outlined and might be summarized as follows:
(a) The approach is formalist in spirit; that is, it attempts to formulate its
theories in terms of symbolic mechanisms of the sort that have
grown out of symbolic logic.
(b) The level of analysis, or the level at which the explanations or theo¬
ries are cast, is functional, and they are described in terms of their
information flow. What this means in particular is that this approach
factors out such questions as how biological material carries out the
function and how biochemical and biophysical laws operate to pro¬
duce the required information-processing function. This factoriza¬
tion is analogous to separating electrical-engineering considerations
from programming considerations in computer science. This does
not mean that such questions of biological realization are treated as
any less important, only that they represent a distinct and, to a large
extent, independent area of study.
(c) In addition to factoring out organic questions, this approach is also
characterized by the techniques it uses in formulating its theories
and exploring the consequences of its assumptions. The most widely
used (though not universal) technique is that of computer implemen¬
tation. Thus, an important methodological goal of cognitive science
is to specify symbolic mechanisms that can actually exhibit aspects
of the behavior being modeled. Adherence to such a sufficiency
criterion makes this approach in many respects like a design disci¬
pline rather than a natural science, at least insofar as the latter typi¬
cally attempts to uncover a small set of fundamental axioms or laws.
Its concern with synthesis makes it, to use Herbert Simon’s phrase,
one of the “sciences of the artificial.’’ [Simon, 1969.] By this, Simon
means that the search for ways in which a certain task can be carried
out within specified constraints—for example, of time and space—is
the mark of a design discipline. Of course, in the case of cognitive
science many of the constraints originate with empirical observa¬
tions. These observations may be not only of what tasks people can
carry out, but also of such things as the time it takes people to solve
certain kinds of problems, which ones they find most difficult, and so
on. Nonetheless, adopting the criterion of constructing a working
system as part of the explanatory process makes the cognitive
INFORMATION SCIENCE VIEWED FROM COGNITIVE SCIENCE 69
the other hand, these indicators pertain to the general population of publica¬
tions: They say nothing about how the most important and most influential
work is being done. Much of the most ground-breaking work is, in my view,
clearly cross-disciplinary and has been extremely influential in shaping peo¬
ple’s thinking about cognitive science. The reason that this does not show up
in statistical patterns of citations is that the number of such seminal works is
(still) small. As an example of this, the aggregate statistics presented do not
reveal that some important cross-disciplinary works' form the bulk of the
small set of cross-disciplinary citations in both psychology and artificial-
intelligence literature. Thus, while the total number of cross-disciplinary
citations may not be large in relation to the total number of publications, a
few of the clearly cross-disciplinary works have been extremely influential.
And that is perhaps to be expected in an area that purports to form a new
discipline as opposed to just being an amalgam of existing fields.
The same pattern emerges when one examines various other possible
indicators of the relation between cognitive science and its contributing
disciplines. For example, membership in various professional societies and
on editorial boards of journals indicates that most people who belong to
these official bodies belong to the traditional disciplines but that, as in the
case of the citation statistics examined, there are a small number of highly
influential cross-disciplinary members. Also paralleling the finding for cita¬
tions, both the editorial board of Cognitive Science and the governing body
of the Cognitive Science Society contain a broad distribution of members
from the contributing fields, especially psychology, computer science, and
linguistics.
I have examined historical relations among fields that have recently come
together as cognitive science and have briefly looked at several indexes of
their de facto interrelation in the practice of research in this new field.
However, the practitioners of cognitive science have a general view of what
disciplines currently house the problems that are at the core of cognitive
science. These include such fields as anthropology, computer science, lin¬
guistics, neuroscience, philosophy, and psychology. They also include such
existing cross-disciplinary areas as cybernetics, neurolinguistics, neuro¬
psychology, computational psychology, computational linguistics, philoso¬
phy of psychology, philosophy of language, anthropological linguistics, cog¬
nitive anthropology, and brain evolution. A committee appointed by the
‘Prime examples are Newell and Simon’s Human Problem Solving [1972]; Winograd’s Under¬
standing Natural Language [1972al; Ernst and Newell s GPS: A Case Study in Genetality
[1969]; and such collections as Bobrow and Collins’s interdisciplinary reader [1975],
74 ZENON W. PYLYSHYN
Board of the Alfred P. Sloan Foundation (to advise the foundation on fruitful
areas of frontier research deserving special support) recently put together an
informal internal report on the state-of-the-art in cognitive science.2
The introductory section of this report mentions the preceding areas of
scholarship as being the contributing fields of cognitive science, while warn¬
ing that a general integration of the cognitive elements of these distinct
disciplines is not yet at hand: Cognitive science is still a goal rather than a
reality. However, it is a goal that an increasing number of scholars endorse
and has led in the last few years to the creation of several dozen centers of
excellence devoted to its pursuit.3 The portion of the Sloan Report included
in the appendix describes in general terms the current vision of the scope of
this discipline.
An inspection of the general picture just presented and in the appendix
reveals that cognitive science is not identical with what is commonly re¬
ferred to as information science, inasmuch as it does not mention many
fields that are concerned with storing, processing, and communicating infor¬
mation—such as work on information theory (including theories of optimal
encoding, cryptography, and information security); information retrieval
from large data banks; practical work in network development or library and
document-management systems; and many other such research areas that
are intimately involved with studying information. None of these studies can
be excluded as irrelevant to cognitive science, since any of them could
produce insights on the technical aspects of information use in cognitive
(i.e., thinking) systems. Nonetheless, these studies are not primarily di¬
rected at understanding the nature of human and nonhuman cognition or the
exercise of intelligence in the course of relating systems to their environ¬
ments. Thus, it is fair to view cognitive science as primarily an empirical
natural science concerned with a subset of the class of problems in informa¬
tion science, namely, those that bear on the general question of the exercise
of intelligence by systems that exist at least partially in an autonomous
relation with a natural and social environment to which they are actively
adapting.
2 Authors of this report included Michael Arbib, Carl Baker, Joan Bresnan, Roy D’Andrade,
Ronald Kaplan, Jay Keyser, George Miller, Donald Norman, Zenon Pylyshyn, Scott Soames,
Richard Thompson, Edward Walker, and Edgar Zurif.
3 Among the North American universities that have established centers for cognitive science are
Brown University, University of California at Berkeley, Carnegie-Mellon University, Univer¬
sity of Chicago, University of Colorado, Cornell Medical School, University of Illinois, Univer¬
sity of California at Irvine, University of Massachusetts, Massachusetts Institute of Technol¬
ogy, University of Michigan, University of Pennsylvania, University of Pittsburgh, Princeton
University, University of Rochester, Rutgers University, University of California at Santa
Barbara, University of California at San Diego, Stanford University, University of Texas,
University of Western Ontario, and Yale University. In addition, a large number of industrial
and other research institutions have also developed cognitive science laboratories—among
many others, Bell Laboratories, Bolt, Beranek and Newman, Exxon, Fairchild, IBM, Rand,
Schlumberger-Doll, Stanford Research Institute, and Xerox.
APPENDIX 75
Appendix
COGNITIVE SCIENCE, 1978
Report of the State of the Art Committee to the
Advisors of the Alfred P. Sloan Foundation,
October 1, 1978
Introduction
Philosophy
The samples for which people have reliably communicable color names
(the sort of names which would enable a second person to select the color on
the basis of the name assigned) were found to be just those that could be
recognized and selected from others after a time interval. Colors without
simple names were both difficult to name or describe and difficult to remem¬
ber.
Such results initially were interpreted to mean that language does con¬
strain thought—in this case, that habits of color-naming determined the
accuracy of color memory. This interpretation proved to be inverted. Subse¬
quent studies demonstrated conclusively that particular colors are not dis¬
tinctive because a culture has given them names. Rather, the colors are
given names because they are distinctive.
The evidence comes from several sources. Psychological studies revealed
that the primary colors—those which English speakers call red, green, yel¬
low, blue, black, and white—have special status in color discrimination
tests. Neurophysiological research demonstrated that the visual system or¬
ganizes these primaries into opponent pairs—red/green, yellow/blue, and
black/white. In other words, the organization of color into primary pairs is
an innately given biological property of all humans with normal vision. Fur¬
thermore, anthropological studies of peoples whose languages contain fewer
color terms than English revealed that, in all cases, the “best instances” of
the colors named are close to the psychologically primary colors; that these
color terms are directly translatable to English. Finally, when memory tests
were conducted on people who spoke a language with only two basic color
terms, they remembered primary colors best, even though they had no
names for them.
Clearly, some aspects of color-naming are culturally determined: the
number of terms does vary. However, the use made of available color ter¬
minology is constrained by innate neurophysiological properties and their
relation to cognition rather than vice versa. In short, the conceptual capacity
underlying color-naming conforms more closely to the universalist than to
the relativist position.
Of itself, this conclusion is of no momentous significance: color terms
form only a fragment of the lexicon in any language. Our present interest,
however, is to demonstrate the kind of interdisciplinary research strategy
that is required to establish such a result.
Note first that the intellectual significance of the color-naming studies
derives from a general theoretical question about the nature of the human
mind, a question which is not properly restricted to any one of the disci¬
plines that contributed to the research just described. We assume that this
one research objective of cognitive science is to formulate abstract descrip¬
tions of the mental capacities manifested by the structure, content, and
function of various cognitive systems. We will refer to the subject matter of
this kind of objective as abstraction.
Second, note that the research on color-naming would have been impossi-
APPENDIX 79
perhaps significant, change. As the field matures, there is little doubt that
great progress will be made in many of the aspects of cognition currently
under investigation. Perhaps the single most important contribution which
can be made to further this effort would be to introduce centers of research
in which active scholars and students might come together for extended
collaboration in research and training at all levels. This report is intended to
demonstrate the value of, and immediate need for, such a contribution to the
field.
COGNITIVE SCIENCE
The View from Brain Theory
Michael A. Arbib
This paper refers to research supported in part by NIH grant NS 14971 and by A. P. Sloan
Foundation grant 80-6-13 to the University of Massachusetts. Portions of this paper appeared in
somewhat different form in my article in the Handbook of Physiology—The Nervous System II.
Motor Control [1981],
81
82 MICHAEL A. ARBIB
beyond the scope of the present commentary to trace the further evolution
of work under the banner of cybernetics. Rather, let us simply note that, as
the field developed in the fifties, it began to fragment. Much work in cyber¬
netics now deals with control problems in diverse fields of engineering,
economics, and other social sciences, whereas the broad field of computer
science has become a discipline in its own right. Here, we briefly cite five
subdisciplines that have crystallized from the earlier concern with the inte¬
grated study of mind, brain, and machine.
In the 1970s, a new grouping took place, which brought together research¬
ers in artificial intelligence and cognitive psychology with those linguists and
philosophers of mind who emphasize symbol-processing. The resulting field
84 MICHAEL A. ARBIB
A notable characteristic of the brain is the orderly mapping from one neural
layer to another, be it the retinotopic mapping from the retina to the many
visual systems or the somatotopic mapping of the motor cortex to the mus¬
culature. I briefly look at one hypothesis of how such a map acts as a control
surface such that the spatio-temporal patterns in the map provide input to
some control system in the brain. Pitts and McCulloch offered a distributed-
processing model of the reflex arc that extended from the eyes through the
superior colliculus to the oculomotor nuclei, thereby controlling muscles
that direct the gaze so as to bring the fixation point to the center of gravity of
distribution of the visual input’s brightness. [Pitts and McCulloch, 1947.]
Pitts and McCulloch noted that excitation at a point on the left colliculus
corresponds to excitation from the right half of the visual field and so should
induce movement of the eye to the right; gaze is centered when excitation
from the left is exactly balanced by excitation from the right. Their model is
so arranged, for example, that each motoneuron controlling muscle fibers in
the muscles that contract to move the eyeballs to the right should receive
excitation summing the level of activity in a thin transverse strip of the left
colliculus. This process provides all the excitation to muscles turning the eye
to the right. Reciprocal inhibition by axonal collaterals from the nuclei of the
antagonist eye muscles, which are excited similarly by the other colliculus,
performs subtraction. The quasicenter of gravity’s vertical coordinate is
computed similarly. Eye movement ceases when and only when the fixation
point is the center of gravity.
This scheme shows how to design a retinotopically organized network in
which there is no “executive neuron” that decrees which way the overall
system behaves; rather, the dynamics of the effectors, with assistance from
neuronal interactions, extract the output trajectory from a population of
COGNITIVE SCIENCE: THE VIEW FROM BRAIN THEORY 85
neurons, none of which has more than local information as to how the
system should behave. In other words, the Pitts and McCulloch model of the
superior colliculus shows how the organism can be committed to an overall
action by a population of neurons, none of which has global information
about which action is appropriate.
I have argued that the study of such cooperative computation in so-
matotopically organized networks provides a central paradigm in brain
theory. [Arbib, 1972.] The study of visuomotor coordination in frog and toad
provides an area (by no means the only one) rich in experiments that may
contribute to our understanding of the neural underpinnings of perceptual
structures and distributed motor control. A number of experiments on prey
catching and predator avoidance as well as on the modification of behavior
by the presence of barriers have led to models of prey selection and prey-
enemy pattern-recognition. [Ewert, 1980; Arbib, 1982.] These experiments
provide support for the following general conclusions:
1. Analysis of vision must not only study local features and responses to
point stimuli, but also develop techniques to analyze structured visual
stimuli. Visual processing can be viewed in terms of cooperative computa¬
tion in somatotopically-organized networks, with overall patterns of activity
being generated in neural structures wherein no single neuron has global
information as to what course is appropriate. Here, there is a rich area of
relevant artificial intelligence totally ignored in Pylyshyn’s review, namely,
machine vision. For example, Dana Ballard and Christopher Brown ap¬
proach machine vision in a way that is increasingly based on a brainlike style
of cooperative computation. [Ballard and Brown, 1982.] As discussed in my
“Cybernetics: The View from Brain Theory” in this volume, there is now
developing a general theory of competition and cooperation in neural nets,
with applications to prey selection, mode selection, and stereopsis, and with
similarities to relaxation techniques used in artificial intelligence. An impor¬
tant bridge between machine vision and human vision is offered by Marr.
[Marr, 1982.]
2. It is useful to view a neural map as a control surface providing input
to some control system in the brain. In addition to a knowledge of competi¬
tion and cooperation within the neural nets that constitute each controller,
there must be an understanding of the cooperative computation of control¬
lers. This cooperation may entail coordination turning off all but one of the
controllers, or it may involve a rich interplay between subsystems, each
modulating the other.
words are: “questions of biological realization are [not] treated as any less
important, only . . . they represent a distinct and, to a large extent, indepen¬
dent area of study.”) In fact, theories apply blinkers to the consideration of
relevant facts, and a cognitive scientist who restricts himself or herself to a
symbol-manipulation theory will never examine the data on brain function
that could show the inadequacy of this approach. Pylyshyn emphasizes the
top-down analysis of much cognitive science. However, he has elsewhere
given a more subtle analysis of the way in which the data the scientist
chooses to examine determine the appropriate grain of the functional ar¬
chitecture of the theory. [Pylyshyn, 1980.] Just consider the very different
demands of linguistic models that address the following three kinds of data:
judgments of sentence grammaticality, response time and eye-movement
data on sentence comprehension, and aphasiological data on the effect of
brain damage in language performance.
In the rest of this comment, I want to show how an analysis of perception
and motor control in the style of top-down cognitive psychology can fruit¬
fully interact with bottom-up analysis enriched by the analog (rather than
symbolic) control theory of cybernetics. The approach grows out of my
concern with brain theory, and is developed in more detail in my contribu¬
tion to the Handbook of Physiology. [Arbib, 1981.] However, the neural
aspects of the analysis will be played down in the present discussion.
ment, anticipating objects before they come into view. The information
gathered during ego motion must be systematically integrated into the organ¬
ism’s internal model of the world: Information picked up modifies the per-
ceiver’s anticipations of certain kinds of information that, thus modified,
direct further exploration and prepare the perceiver for more information.
Such considerations will, 1 predict, increasingly engage the attention of cog¬
nitive scientists as robotics becomes an increasingly active area of study
within artificial intelligence. The problem of controlling robot arms and in¬
tegrating visual and tactile information will require increasing attention to
control theory, cooperative computation, and sensorimotor integration. This
will combat the overemphasis on symbolic processes provided by
Pylyshyn’s paper.
The intelligent organism does not so much respond to stimuli as select
information that helps it achieve current goals, although a well-designed or
evolved system certainly needs to take appropriate account of unexpected
changes in its environment. Planning is the process whereby the system
combines an array of relevant knowledge to determine a course of action
suited to current goals. In its fullest subtlety, planning involves the refine¬
ment of knowledge structures and goal structures as well as action per se.
Novel inputs (e.g., encountering an unexpected obstacle) can alter the
elaboration of high-level structures into lower-level tests and actions that, in
turn, call on the interaction of motor and sensory systems. Scientists seek to
study programs that are part of the internal state of the system and can
flexibly guide ongoing action in terms of internal goals or drives and external
circumstances. My thesis is that perception of an object (activating appropri¬
ate perceptual schemas) involves gaining access to routines for interaction
with the object (motor schemas) but does not necessarily involve executing
even one of these routines. Although an animal may perceive many aspects
of its environment, only a few of these can at any time become primary
loci of interaction. Therefore, perception activates (i.e., defines a search
space, draws a map), and planning concentrates (lays out the route to be fol¬
lowed).
I use the term perceptual schema to denote the process whereby the
system determines whether a given domain of interaction is present in
the environment. The state of activation of the schema then determines the
credibility of the hypothesis that what the schema represents is, indeed,
present, whereas other schema parameters represent such properties as
size, location, and motion of the perceived object. Consider a schema that
represents, say, a chair; also consider an environment that has two chairs in
plain view. It is clear that two copies of the chair schema (or at least two
separate sets of chair-schema parameters) are required to represent the two
chairs. These two copies are separate instantiations of the same schema,
each with its own set of parameter values. The internal representation of the
environment may thus be viewed as an assemblage of spatially tagged,
parametrized schema-instantiations.
88 MICHAEL A. ARBIB
Motor Schemas
I have used schema to indicate the type of unit from which the internal
representation of the environment can be built. Programs for motor control
may themselves be seen as assembled from suitable units, which are referred
to as motor schemas. These motor schemas are related to synergies (in the
sense of the Russian school founded by Nikolai Bernstein [Bernstein, 1967]).
These schemas may exhibit short-term memory. To control properly the
motion of an object (the controlled system), the controller must clearly know
such relevant parameters of the object as its mass and moment of inertia.
However, in the real world the exact values of these parameters are seldom
available and may actually change over time. (Compare short-term loading
effects on muscles with longer term aging effects and weight changes.) To
adapt to such changes, the feedback loop must be augmented by an
identification algorithm. The job of this algorithm is to identify more accu¬
rately the parameters of the controlled system. To do this, it continually
monitors the behavior of the controlled system and compares it with the
output that would be expected on the basis of the current estimated parame¬
ters. Any discrepancies in the output can be used to obtain more accurate
estimates of the parameters that define the controlled system. These updated
parameters can then be supplied to the controller as the basis for its state
estimation and control computations.
Note well that the identification algorithm can do its job only if the con¬
troller is of the right general class. It is unlikely that a controller adapted for
guiding the arm during ball-catching would be able, simply as a result of
parameter adjustment, to control properly the legs in performing a waltz.
Thus, the adaptive controller (controller plus identification procedure) is not
to be thought of as a model of the brain; rather, each such control system is a
model of a single motor schema that can be activated when appropriate.
This framework for analyzing visually guided behavior of a complex or¬
ganism is based on four general premises:
Karl Lashley laid many of the foundations for neuropsychology, that most
cognitive of the neurosciences. In an important critique of stimulus/response
theory, he raised questions about serial order in behavior that are answered
at the conceptual level as soon as one thinks of the brain’s computations not
in terms of stimulus/response couplings or chains of associations but, rather,
in terms of coordinated control programs. [Lashley, 1951.] Although our
knowledge of computer programs removes the conceptual problem of serial
order, the question of how such control strategies can be neurally imple¬
mented is only beginning to be answered. Much of the neurophysiological
analysis of movement has focused on spinal mechanisms (especially feed¬
back mechanisms in posture and locomotion) and on higher level single-cell
correlates of stimulus or response. Future research clearly must aim to
better analyze the distribution of planning operations within cortical struc¬
tures and understand the signal flow this planning must impose on the cere¬
bellum and other regions that modulate this planning.
Even though the neural mechanisms for the planned coordinated control
of motor schemas seem to be beyond the range of current experimental
investigation, I suggest that artificial-intelligence approaches to planning
may provide a framework for the development of such investigations in the
future.
A pattern of action may be quite complex, with the actions intertwined
and overlapping. Simultaneous actions must be coordinated, and successive
actions must be smoothly phased, one into the next. In this section, I discuss
the concept of a coordinated control program as the type of structure that
orchestrates the interwoven activation of motor schemas controlling differ¬
ent actions.
Biological control theory usually studies neural circuitry specialized for
the control of a specific function, be it the stretch reflex or the vestibulo-
ocular reflex. Yet, most behavior involves complex sequences of coor¬
dinated activity of a number of control systems. Thus, I explore the notion
of a coordinated control program as a combination of control theory and the
computer scientist’s notion of a program suited to analyzing the control of
movement. Control theorists use a block diagram to represent a system.
Each box represents a subsystem that is continually active, whereas the
lines linking the boxes illustrate the transfer of data, showing how the output
90 MICHAEL A. ARBIB
CONCLUSION
Michael S. Gazzaniga
Zenon Pylyshyn has laid out with striking clarity what cognitive scientists
do, which is, mainly, test the design of information systems that model
aspects of a behavior. That is not what cognitive neuroscientists do, and my
bet is that the approach he describes will be inefficient in designating cogni¬
tive systems that do accurately describe how human minds work. That is the
task cognitive neuroscientists have set for themselves, and their objective is
to take up the problem experimental psychologists left for neuroscientists
but that neuroscientists did not pick up. This is because the current view in
neuroscience is that the cellular study of brain function will yield more
substantial insights into, well, into something, and the question is what is the
something? Accepting completely Pylyshyn’s case for the levels-of-analysis
approach, cognitive neuroscientists believe that the cellular approach of
their colleagues in basic neuroscience will tell us about cells. That is an
admirable task, but one that leaves begging the task of understanding human
cognition.
In order to compare the approaches, it is helpful to take the example of
the problem of memory. The basic neuroscientists, taking the lead from
work on brain-damaged humans, have decided to focus a large part of their
energies on a brain structure called the hippocampus. This structure, which
is implicated in the human memory process, can be removed from experi¬
mental animals, placed in tissue culture, and the metabolic and electrophysio-
logical properties of its cells studied. A series of elegant reports has demon¬
strated that changes in cellular activity occur to repetitive stimulation that
are long lasting, indeed, seemingly permanent. The biochemical machinery
responsible for this response seems to be in hand and investigations are well
on their way to determining the site of morphological changes on the
synapse that must occur to explain the long-lasting effect. Nice, perhaps
even brilliant, but how such changes connect with how a person remembers
a telephone number when it is breezed by at a cocktail party is not even
attempted.
At the level of description at which cognitive scientists work, the task is
to describe how artifacts and/or humans process information. Over the
93
94 MICHAEL S. GAZZANICA
years, cognitive scientists have established limits on the amount and kind of
information that can be apprehended, on strategies that facilitate the acquisi¬
tion of information, and on the possible different kinds of memory the cogni¬
tive system handles. These elegant studies generate data on response char¬
acteristics of these human and nonhuman information systems, and theories
are constructed about how the system must be organized. What is left open,
of course, is whether or not the human brain is organized the way the
cognitive scientists think it must be, just as Pylyshyn plainly says.
Alas, the cognitive neuroscientists’ approach! The subject for experimen¬
tal observation is the human with brain disease and/or focal lesion that can
be accurately localized and described. The measures are on this human’s
ability to perform memory tasks of a wide variety. The results allow one to
construct theories about how the memory system is organized by consider¬
ing how systems make errors that are reliably generated by specific patho¬
logical states. The approach puts constraints on the theories of the cognitive
scientists and the neuroscientists. At the present time, the cognitive neuro¬
scientist is rewriting the backdrop for researchers in both areas. The neuro¬
scientist is being redirected to other brain areas as well as facing the reality
that more than one cellular process may be responsible for memory. Cogni¬
tive scientists are having to adjust their theories to the fact that there is not
one cognitive system that is in charge of all information-processing. Recent
developments would suggest that the cognitive system has a modular-type
organization and that these modules may have specialized features.
Of course, in reality, all of the approaches seem terribly exciting, and any
contemporary scientist must become reasonably conversant in the three
fields—cognitive science, basic neuroscience, and cognitive neuroscience—
but each of us, happily, has our own preference.
COGNITIVE SCIENCE
A New Patch in the Quilt?
Saul Gorn
The very fact, however, that they have symbolisms with well-defined
manipulative procedures may make it useful for them to simulate their pro¬
cesses symbolically. It is one of the informatician’s basic insights (prejudices,
superstitions, ideologic attitudes, or peculiar ways of looking at the world?)
COGNITIVE SCIENCE: A NEW PATCH IN THE QUILT? 97
that any process that can be precisely specified is capable of being symbolically
simulated .... (From my paper in this volume.)
I believe this is not much different from what Pylyshyn says in the second
section of his paper. However, I find his item (1) in that section, the attribu¬
tion of formalist properties (as though it were mathematical) puzzling; for
example, artificial intelligence is much more empirical than formal.
In this connection, I have a question to ask Pylyshyn the psychologist
rather than Pylyshyn the computer scientist: Are not the semantics of the
symbols representing the basic cognitions of most disciplines definitely not
symbol-manipulative? Do they not refer to what they consider to be real
things and not symbols? Is it not only the informatician whose basic con¬
cepts are about symbol manipulation? Even mathematicians resent being
confused with arithmeticians or accountants. How, then, should I interpret
this sentence from Pylyshyn’s paper: “In all these instances, the behavior of
the systems in question (be they minds, computers, or social systems) is
explained, not in terms of intrinsic properties of the system itself, but in
terms of rules and processes that operate on representations of extrinsic
things.” Is this not a computer scientist talking, rather than a psychologist
or a cognitive scientist? Perhaps the title of Pylyshyn’s paper should have
been reversed to read “Cognitive Science: Its Roots and Relations as
Viewed from the Perspective of Information Science.”
Finally, a remark on Pylyshyn’s analysis of the citation statistics is in
order. I believe he uses them as evidence in response to his question about
convergence into one discipline. Although I go along with him on the conver¬
gence question, I do not think citation indices form conclusive evidence.
When a new disciplinary area is being formed, one would naturally expect a
transition period during which the fusion process accompanies the fission
process by which the new body is separated from its parents. It is precisely
when there is a Kuhnian scientific revolution that the new generation breaks
contact with the old (which continues to follow its same old paradigms). Is
this not the standard way in which a new patch is formed in the patchwork
quilt of human lore? In this case, the corners of eight abutting patches (the
previously mentioned cognitive disciplines) are cut off and fused to form a
new patch. If, then, a convergence is occurring (and, I repeat, I am skept¬
ical), present statistics would still be not only what is expected, but would
most likely be misused by the old guard as an argument against the young
Turks.
.
REFLECTIONS ON THE
STRUCTURE OF
AN INTERDISCIPLINE
Allen Newell
In his “Information Science: Its Roots and Relations as Viewed from the
Perspective of Cognitive Science,” Zenon Pylyshyn outlines the develop¬
ment of cognitive science, an interdiscipline that lies within something called
information science, this latter being too amorphous and wide-ranging to be
termed even an interdiscipline. Here follow a few thoughts on the nature of
interdisciplines.
COMPLICATIONS
This research was sponsored in part by the Defense Advanced Research Projects Agency
(DOD), ARPA Order No. 3597, monitored by the Air Force Avionics Laboratory under Con¬
tract F33615-78-C-1551. The views and conclusions contained in the paper are those of the
author and should not be interpreted as representing the official policies, either expressed or
implied, of the Defense Advanced Research Projects Agency or the United States government.
99
100 ALLEN NEWELL
potential path, if a scientist does not know how to do the next piece of
research down that path, he or she will rarely go that way.
The point is that modern cognitive psychology is as much a product of the
earlier history of psychology and developments internal to cognitive psy¬
chology, as it is of the external ideas of information-processing. In turn, this
cognitive psychology is a major determiner of the character of the interdisci¬
pline of cognitive science and cannot be left out of the picture.
The story of linguistics, even more than psychology, seems to me one of
internal determination, although I am on less sure ground here. To an out¬
sider, the Chomskian revolution arises from the same general intellectual
milieu that Pylyshyn sketches, impacting linguistics rather than psychology.
The algebraic character of the new linguistics and the fundamental idea of
generative rules characterizing behaving systems—these seem a product of
the same intellectual viewpoint of operational mathematics that gave rise to
operations research, game theory, and control theory. But to an insider,
perhaps to Chomsky himself, the development of the new linguistics looks
almost entirely internal. When he traces roots, as in his Cartesian Linguis¬
tics, it is in terms of internal linguistic history and such general philosophical
positions as rationalism and empiricism. [Chomsky, 1966.] It is certainly not
the Descartes of analytical geometry that the book’s title refers to.
Again, from an internal linguistics perspective, I would suspect that the
continued development would be seen mostly as flowing outward. Certainly,
one can document a transformation of psycholinguistics in the sixties that
flowed directly from linguistics, but left linguistics only modestly changed, if
at all. And there has remained a barrier, both intellectual and political,
between linguistics and computer science, which has kept computational
linguistics in a rather ambiguous status. My concern here is not primarily to
assess or evaluate such relations, but only to note how they complicate the
already complex picture that Pylyshyn sketches.
Another kind of complexity exists, which can be illustrated by the case of
artificial intelligence. Here we have a field that has a fair amount of coher¬
ence, but only a short time ago (the sixties) was itself considered an interdis¬
cipline. How come it is now seen as a disciplinary pillar on which one can
build a new interdiscipline of cognitive science? Indeed artificial intelligence
is not a discipline, in the usual sense of the word, but rather a part of
computer science—which itself is newly come to disciplinary status. In their
wisdom, the makers of the Sloan figure did put computer science as a node
rather than artificial intelligence. But the reality is that almost the entire
connection with computer science lies with artificial intelligence. As an indi¬
cation of how jerry-rigged the structure is, artificial intelligence did not get
itself its own national professional society—the American Association
for Artificial Intelligence—until after the Cognitive Science Society was
formed.
The complexities continue when we consider neuroscience. It also is an
interdiscipline—the neuro parts of anatomy, physiology, pharmacology, and
ALLEN NEWELL
102
Semiotics is the study of signs, i.e., of those entities which effect communica¬
tion between interpreters of signs. Quite a variety of things can function as
signs. A word, a sentence, a gesture, a facial expression, a photograph, a
diagram, etc., are all signs because we, their interpreters, are more concerned
with what they stand for or represent than with what they are merely in them¬
selves. (Laferriere, 1979, p. 434.)
Perhaps the difficulty is that we are casting the story in terms of the disciplin¬
ary structure of science. There are certainly alternative ways to tell the story
of an intellectual domain such as information science or cognitive science.
Perhaps the situation would not seem so complicated, if we dealt directly
ALLEN NEWELL
104
WHAT IS A DISCIPLINE?
defining the intellectual ground against which scientists do their work; that
is, pursue their specialties.
Thus, a significant aspect of artificial intelligence not being a discipline in
itself and being part of the discipline of computer science is what the scien¬
tist in artificial intelligence knows about besides his or her own research,
say, on expert systems. If artificial intelligence were a discipline, all the
scientist would know about are other aspects of intelligent systems, such as
vision, theorem-proving, and natural language. With artificial intelligence
part of computer science, the scientist also knows about work on multipro¬
cessing architectures, verification, algorithm analysis, and so on. Corre¬
spondingly, of course, he or she knows only a little about what is going on in
cognitive psychology and nothing at all about what is going on in social
psychology.
If, by some chance of fate or fashion, artificial intelligence had become
part of psychology or philosophy or, say, a developed cybernetics, then
quite different things would be part of the ground. Advances in behavior
modification, phenomenological philosophy, or muscle control, respec¬
tively, would be a living part of the artificial intelligence scientist’s world.
Correspondingly, VLSI (very-large-scale integration) and the structure of
protection systems would not.
What intellectual ground a scientist stands on is immensely important. It
affects what intellectual resources are drawn on in choosing frameworks for
analysis, interpretation of difficulties, sources of inspiration, and the kernel
ideas for solutions to conceptual problems. It affects to whom the scientist
can talk about his or her problems—namely, those who share common
ground—and, therefore, what types of suggestions he or she obtains. In
short, though it does not affect the position or velocity of the work, which is
determined by the research task itself, the discipline affects all higher de¬
rivatives.
One fundamental cognitive reason why disciplines exist seems appar¬
ent—it comes from resource limitations to human cognition. There are only
so many hours in the day; they can be filled with only so much talk and only
so many articles. These limits assure that no scientist will be familiar with
more than a small fraction of science; thus, science cannot be the seamless
web we all desire. However, such limits do not determine that science will
partition itself into (disciplinary) clumps, in which many scientists read ap¬
proximately the same scientific literature and few read heavily across disci¬
plines. It seems equally plausible that each scientist would trace out an
idiosyncratic problem-oriented trajectory through science. The whole of sci¬
ence would then have the appearance of a randomly woven but homoge¬
neous-appearing mat—a crazy quilt at the individual level rather than at the
discipline level.
There is an entirely different side to disciplines and their structure,
namely, they are social institutions and serve social and motivational needs.
They provide the rocks to which scientists attach their identities. “Hello,
ALLEN NEWELL
106
Given the disciplinary structure, as just described, what then is the role (or
set of roles) of interdisciplines? There are two obvious ones. First, science is
in fact all of a piece; thus, any decomposition into disciplines must necessar¬
ily be imperfect. Interdisciplines represent a spillover as groups of scientists
with common interests read each other’s work and begin to cooperate with
each other. In this role, interdisciplines can be permanent and stable institu¬
tional structures in science.
Second, the division into disciplines changes with history. Early in the
development of science, new sciences precipitate out of a less differentiated
scientific and intellectual activity—for example, natural philosophy. The
emergence of physics and chemistry as separate disciplines would be an
example, or the emergence of statistics. As science matures, all of the terri¬
tory is covered in one way or another by some science, then new sciences
are carved out of existing ones. Interdisciplines are simply early forms of
disciplines on the way to existence; biochemistry provides an example.
Other roles are possible: An interdiscipline could be an intermediate stage
where part of one discipline moves to another. It could be a temporary
scaffolding while some important ideas are transmitted to a set of fields, only
STRUCTURE OF AN INTERDISCIPLINE 107
I end where I began. The disciplinary picture for cognitive science is a crazy
quilt and even more so for information science. I believe it will stay that way
almost indefinitely. The intellectual domain of understanding human nature
is too extended and admits too many approaches to enforce unification. All
that is required to avoid any serious pressure for unification is that en¬
deavors be separate in their problems. Overlapping claims need not violate
this, because claims are made by area, but separation is determined by
problem. As long as the approaches in a field that answer one problem do not
suffice to answer nearby ones, then many approaches can coexist with little
pressure to unify them. The progress of an approach taken in isolation is not
relevant—it can be real, measurable, and even dramatic, without implying
that other approaches to the same arena should be abandoned.
Will it always be this way? That is too far in the future to guess. I believe
the essential condition for the gradual simplification of the disciplinary pic¬
ture for cognitive science lies in the development of techniques that solve
well-specified scientific problems and answer scientific questions in routine
ways. The phonetic alphabet is a good example; we all depend on linguistics
for that. The representations for grammars are another; I believe they will be
around and used routinely by all scientists, even when our conceptions of
language have evolved considerably. Similarly, from psychology comes sig¬
nal-detection theory, with the concept of the operating characteristic, and
from artificial intelligence comes the LISP programming language. These all
become islands of certainty that all disciplines borrow and use when needed.
Though sometimes such techniques become irrelevant and drift away, by
and large they form a permanent accumulation. They can be expected to
increase gradually (and sometimes rapidly) to form a framework that will
ultimately force simplification of the disciplinary structure. Until that time,
there is no reason for those professing an interdiscipline not to let the crazy
quilt continue. Furthermore, there are good reasons to be confident that
ideas will continue to move from discipline to discipline.
.
■
IN FORMA VORES
George A. Miller
Both the human tendency to think analogically and the dangers of reasoning
by analogy are familiar themes to students of the history of ideas. One
cannot help but wonder what about the human mind compels it to flirt so
outrageously with the potentially disastrous consequences of false analo¬
gies.
The general pattern of analogical thinking, of course, is to explain some¬
thing that is poorly understood by noting its similarity to something that is
well understood. Thus, it is not uncommon to explain complicated natural
phenomena by analogy to human artifacts, since people generally feel that
they understand reasonably well those things they themselves have actually
manufactured. In the life sciences, one thinks of such examples as ex¬
plaining the motions of the limbs by analogy with levers, explaining the
circulation of the blood in terms of the analogy of the heart to a pump,
explaining eyes and ears by analogy to cameras and microphones, or ex¬
plaining metabolism by analogy to heat engines.
Attempts to understand human cognition have generated their own
analogies. The mind has been likened to a cave on whose walls events cast
their shadows, to a slate on which experience can write, to a hydraulic
system for pumping energy into alternative activities, to a telephone switch¬
board that connects ideas to one another or responses to stimuli, even to a
hologram that stores representations distributively. The current favorite, of
course, is the modern, high-speed, serial, stored-program, digital computer,
an analogy that has catalyzed the rebirth of cognitive psychology as an
active scientific enterprise.
In What is Life?, a little book that opened up biology for physicists, Erwin
Schrodinger pointed out that organisms survive by ingesting, not food, not
calories, but negative entropy. [Schrodinger, 1945.] It is no accident, of
course, that the mathematics of entropy are also the mathematics of infor¬
mation. The analogy is obvious: Just as the body survives by ingesting
negative entropy, so the mind survives by ingesting information. In a very
general sense, all higher organisms are informavores.
But how should this suggestive parallel be developed? No sooner was the
question asked than computer scientists offered an answer. The computer
ill
112 GEORGE A. MILLER
same price progress in formal logic was bought at the beginning of the
century—at the price of a total divorce from psychologism.
But the informational level may be richer than we now know. That is to
say, computers as we presently know them may prove to be incomplete
analogues of informavores. The claim would not be that there is something
computable that they cannot compute but simply that evolution may have
produced a species of informavore—a kind of special purpose computer—
that we as yet have no idea how to construct or imitate.
* * *
Zenon W. Pylyshyn
our present notion of the subject matter of such a science is close to being
what philosophers call a natural kind. It has been widely recognized ever
since Plato first drew our attention to it—that in order to develop a success¬
ful explanatory system, we must “carve nature at her joints." The way we
group and distinguish properties and phenomena is crucial to the success of
the scientific project. Thus, for a successful theory of motion, we must not
distinguish, as did the ancients, between natural and violent motion, though
we do have to distinguish between weight and mass, a distinction with which
medieval thinkers struggled but which was not clarified until the seventeenth
century. Such distinctions serve to carve out a natural and relatively homo¬
geneous scientific domain within which a set of explanatory principles apply.
It had been assumed for centuries that the appropriate natural domain of
study for what we informally call cognition (or higher mental functions) is
something like complex biological systems. In the 1940s and 1950s, this
domain was expanded beyond biology to include what were called cyber¬
netic systems. Gradually, and through the influence of the sorts of histoiical
developments to which several of the commentators alluded (most particu¬
larly the work in artificial intelligence), a picture of natural domain began to
evolve. It looked more and more as though the “natural kind to which the
theories would apply were those systems that were governed by repiesenta-
tions or by knowledge and goals. Newell, in a similar analysis of the current
state of the field, referred to this as the “physical symbol system hy¬
pothesis.” [Newell, 19806.]
This was not simply a case of expanding the domain of study by finding
increasingly abstract ways of viewing systems (as has sometimes been
claimed of general systems theory). The representation view also has a
restricting effect. Some of the central concepts of psychology, as George
Miller notes, are excluded in the process. For example, while the distinction
between conscious and unconscious representations is glossed over, the
subjective-feeling component (the “qualia") of mental states falls outside
the domain of such cognitivist explanations. Similarly, the physical forms in
which representational systems are instantiated becomes a separate field of
study. The work of cognitive neuroscientists like Michael Gazzaniga and
Michael Arbib is taken to be a contribution to our understanding of how one
particular class of machines—the biological one—is able to carry out repre¬
sentation-governed processes. Their work is a study of the underlying func¬
tional architecture, not the rule-governed representational processes them¬
selves, since these constitute an autonomous natural domain with
regularities that can be captured independently of particular physical instan¬
tiations. The domain of cognition is thus viewed as a natural kind encom¬
passing all informavores (as George Miller puts it so colorfully) rather than a
human or even a biological category of functioning. What is lost in the
processes is the concept of a domain of study defined in terms of our naive
interests rather than in terms of its scientific tractability. But such losses are
routine in scientific evolutions: Astronomy, too, lost much glamor and won-
REPRESENTATION, COMPUTATION, AND COGNITION 117
der when it was discovered that astrological questions were not in its pur¬
view.
I end with some brief comments prompted by Saul Gorn’s remarks con¬
cerning the notion of a formal symbol and semantics. Being governed by
representations is not the same as being formal and computational. Nonethe¬
less, computation is the only model we have of how a physical system can be
governed by representation. As such, the computation view of mind is
neither a logical necessity nor a metaphor: It is a far-reaching empirical
hypothesis. Thus, Saul Gorn is exactly right when he says that I am viewing
cognition from the perspective of information science rather than the con¬
verse. That is because I believe this is precisely the view that characterizes
cognitive science; it is a sweeping reanalysis of the notion of cognition itself.
What I called the representational metapostulate is not the same as the “basic
insight of informatics” that Gorn refers to; namely, “that any process that can
be precisely specified is capable of being symbolically simulated.” Rather, it
is the much stronger hypothesis that cognition consists in (not “is capable of
being simulated by”) manipulating physically instantiated symbols. It is the
quite literal proposal that cognition is computation.
Because of this postulate, symbols do not have the same status in cogni¬
tive science that they have in either physics or semiotics, though for differ¬
ent reasons in the two cases. In physics, symbols represent physical mag¬
nitudes and the relations among these: They are a notation for the subject
matter only. The symbols themselves are not assumed to have any realiza¬
tion except in the expression of the theory. Thus, the choice of a notation,
though it has important consequences for the utility of the theory, does not
matter for the empirical truth of the theory. In cognitive science, as in
computing, symbols not only refer to some extrinsic domain (say, numbers
or beliefs), but they are also real objects in their own right that the machine
(or, by hypothesis, the mind) manipulates in a manner that preserves the
semantical properties of those symbols. In the case of numerical computa¬
tion, for example, symbol manipulation is carried out in such a way that
when interpreted as numbers according to some uniform scheme, the ex¬
pressions remain congruent with number-theoretic properties. That is why
we can quite literally say that the machine carries out arithmetic processes.
According to the view I have been describing, exactly the same story is to be
told about cognition in general. Because of this, in cognitive science the
notation matters: It constitutes a substantive claim about the cognitive pro¬
cess. Gorn is right in remarking that this is the computer scientist talking,
but that is because computer science and cognitive science converge pre¬
cisely at this point.
Cognitive science is also quite different from semiotics in the kind of
symbols that it studies, which may account for the quite wide divergence
between the two approaches (as Newell notes). Semiotics is concerned with
what might be called secondary symbols. These are the symbols that we
humans use in order to convey meanings. They are not symbols that func-
ZENON PYLYSHYN
118
Saul Gorn
Let me, first of all, choose a shorter expression than computer and informa¬
tion science. Like the French (informatique) and the German (Informatik),
let me choose the word “informatics.” It contains the idea of information
and has an ending like that of mathematics, implying a formally based
theory. It is too bad that informatics loses the computational component in
its name and also leaves no impression of an experimental basis, both of
which I consider equally important, as we shall see.
Previous papers of mine on the subject were addressed to applied mathe¬
maticians, pure mathematicians, library scientists, logicians, linguists,
pedagogues, behavioral scientists, and political scientists. [Gorn, 1965;
1967; 1968.] I even discussed why each audience needed a different presen¬
tation. The audience on the present occasion is the mixture of the neigh¬
boring areas I mentioned first, and the purpose of this paper and those from
the other areas is to help us sort out our separate identities, if they are indeed
separate.
I propose to do my part by first discussing my picture of the disciplines of
human knowledge and action, their ontogenic development, and resulting
121
122 SAUL CORN
cists, mathematicians, and informaticians. (8) How the areas within infor¬
matics will vary from the formal and mathematical, through the empirical
and experimental (especially where unsolvable and unsolved problems are
concerned), through aesthetic and ethical reactions, into the variety of ap¬
plied informatics in a number of professional activities.
Finally, the last section, on the intercommunication and stability of disci¬
plines, will deal with seven topics: (1) The two meanings of ideology that we
are using (as contrasted with the Marxian meaning and its relatives); the
sociology of ideas versus the biology and psychology of ideas and their
connection because of the need of social support for the advanced technol¬
ogy needed to maintain them. (2) How to compare and contrast the ideolo¬
gies of disciplines. (3) Cybernetics as a superdiscipline and robotics as a
professional activity. (4) The managerial-professional activities related to
informatics. (5) Pedagogy as a superdiscipline; the impossibility of a single
pedagogic ideology and the similar impossibility of a single ideology for
information retrieval. (6) The library profession and its long-range stability:
the dependence of the stability of a discipline on both the existence of stable
ideologies (to maintain communication) and the existence of unsolvable
problems (to prevent technological obsolescence). (7) How the stability of
informatics depends on maintaining the interlocking of its scientific, artis¬
tic, and professional activities.
We all know that ever since the agricultural and urban revolutions humans
have tended to specialize their activities. The result is that, by now, human
lore, especially in Western civilization, is a patchwork quilt of areas of
activity called arts, sciences, and professions. In spite of such philosophies
as operationalism, some of these areas still seem to be mainly knowledge-
oriented, such as mathematics, the natural and social sciences, the humani¬
ties, and other academic disciplines, while others seem to be mainly action-
oriented, like farming, sports, hunting, cooking, mining, manufacturing,
constructing, and the like. Still others seem to have as their purpose the
transformation of knowledge into action, such as the professions of en¬
gineering, medicine, law, education, and so forth. Let me use the word
discipline to cover them all, since the dictionary gives the following mean¬
ings: branch of knowledge involving research; rule or system of rules affect¬
ing conduct; and even, training that corrects, molds, strengthens, or perfects.
The patchwork quilt of disciplines is like a series of slightly distorted
plane projections that a mapmaker might make of, say, a doughnut-shaped
planet. Each, therefore, represents a portion of human activity from a
specialized point of view. It seems to me that these specialized representa¬
tions, especially for the more sophisticated disciplines, are for separate ide¬
ologies with definite histories of development and even changes of territory
124 SAUL CORN
We see, then, that it was only when technology had advanced sufficiently
that empiricism could even become conceivable, let alone possible. The
human spectra and ranges of perception and control had to be extensible by
machines, or there would be no evidence on which to base empiricism. The
idea of empiricism may have had early philosophical roots—from Aristotle
to Roger Bacon—but it was not until the seventeenth century that Francis
Bacon’s form of it was expressed, and it took centuries before empiricism
could really flourish.
For example, statics and dynamics developed in the time of Archimedes
(300 B.C.), together with various hydraulic and ballistic technologies. That
was when the concepts of force and weight and pressure outside the human
range really became perceptible and measurable. Similar developments tied
optics to astronomy and biology before measurement could advance into the
macroscopic and microscopic ranges. Certainly, the biological concept of
the cell could have no empirical base before then. It took many decades
before temperature could be measured and centuries before its relation to
mechanics could be appreciated. Thomas Kuhn points out that it is impossi¬
ble to indicate the moment of discovery of oxygen. Over some decades,
Priestley, Lavoisier, Scheele, and others did some individual groping with
experiments and technological devices to control the materials and mutual
communicating on what to capture and how, before it could be said that
anyone knew what oxygen was. [T. S. Kuhn, 1962a; 1977.] It took years for
these scientists to decide what it was that ought to be perceived and to
extend their senses and controls by technological devices in order to per¬
ceive it. This is the social phenomenon that I have elsewhere called a wave
of agreement to perceive. And there had to be this social agreement to
perceive what was unperceivable before, or else the social expense required
to manufacture the artificial effectors and receptors would not be accepted.
This is especially marked in twentieth-century physics and its dependence
on accelerators.
The basic chemical concept of an element, as it was introduced by Dalton
and is thought of today, could not even exist (in spite of the ancient forms of
atomism), let alone have an empirical basis, until after a number of occur¬
rences similar to the story of oxygen; the principle that chemicals combine in
fixed proportions and the principle of conservation of mass was inconceiv¬
able, let alone unmeasurable, until then. Similar things could be said about
the concepts of energy in each of the variety of meanings it has had over the
millenia. The principle of conservation of energy was inconceivable and
inexpressible before the late nineteenth century.
When Thomas Kuhn talks about scientific revolutions, where some of the
paradigms of the science are replaced, I interpret this to mean that the
community forming that science has decided that something considered
worth perceiving before is no longer worth perceiving and is to be replaced
by a new basic concept or a new basic principle, using a new insight. [T. S.
Kuhn, 1962a.] On a millenial time scale, phlogiston disappeared in a flash, to
be replaced by oxygen; on a time scale that could look at day-to-day, week-
126 SAUL CORN
to-week, and month-to-month changes, phlogiston died hard, and the birth
of oxygen was difficult.
Kuhn’s analysis is restricted to those knowledge-oriented disciplines
called the sciences and, more specifically, the empirical sciences and espe¬
cially the physical sciences. Machlup has pointed out that for thousands of
years, the word science referred to the absolutely certain, such as mathe¬
matics and logic and metaphysics but excluded natural philosophy; and that
it was only in the nineteenth century when an empirical base became a
requirement for science that the tables were turned to drive metaphysics
out. [Machlup, 1980.] Machlup points out, however, that in languages other
than English, the word science still includes other disciplines (even law); but
I note that all his examples of the term in other languages are what I have
called knowledge-oriented as distinguished from action-oriented. It seems to
me that even the action-oriented disciplines develop their characteristic con¬
cepts, ideologies, Weltanschauungen, and what I have elsewhere called
mass egos, in a similar way. I will return to this question shortly. Mean¬
while, let me discuss this attitude toward knowledge implied by those con¬
cerned with its accumulation.
Those fond of physical metaphors in psychology and epistemology often
compare knowledge to physical entities in order to invent methods for
weighing and measuring it. They often visualize knowledge as an inert quan¬
tity of stuff independent of the knower. (I believe it was Karl Popper who
criticized and sharpened this point of view. [Popper, 1968.]) Knowledge is
en/Zg/ztenment, and those fond of physical imagery, in talking about it,
should remember that light was not fully understood in the physical sciences
until Huyghens’s wave theory was developed as well as Newton’s particle
theory.
Charles S. Peirce pointed out that the study of signs (symbol systems) was
not merely about a dyadic relation between the symbol and what it means
but was really discussing the triadic relation among symbols, their meanings,
and their users or interpreters. (I will come back to this point when I discuss
the ideology of informatics.) In a similar way, knowledge, which is what we
try to symbolize when we communicate, is not merely a relation between
what is known and the knower; knowledge is the result of an attempt to
communicate and, therefore, involves a community of knowers as well as
what is known. This communication process works like a boundary-value
problem among the knowers of the community, and the things and processes
the community knows in common result from something like a wave phe¬
nomenon. When the concepts of cardinal number, length, area, volume,
mass, weight, temperature, energy, element, or cell, or the processes of
which they partake, were invented (not discovered!), this was a production
of new knowledge that occurred over a period of time. And this social
process was what I called a wave of agreement to perceive. This is the
process that some historians of science call discovery, and whose retraction
and leplacement Thomas Kuhn calls the change of paradigm in a scientific
revolution. [T. S. Kuhn, 1962a.]
INFORMATICS: IDEOLOGY, METHODOLOGY, SOCIOLOGY 127
and clauses or new nouns and verbs of the language. Communication habits
developed by the community make its members build these syntactic types
into sentences and assemblages of sentences.
When the discipline is mainly knowledge-oriented, as in the academic
type, the tendency seems to be to have the majority of the sentences in the
declarative mood; furthermore, the manner of assembling these sentences is
often in deductive or inductive groups, each discipline having its own
method of reasoning or argument. For example, mathematics and physics
use formal and extensional logic, and the assemblages are definitions and
proofs and demonstrations.
When, however, the discipline is mainly action-oriented, the majority of
the sentences tend to be imperative, hence, neither true nor false (though
possibly good or bad). Moreover, the important assemblages are assign¬
ments, allocations, and programs. Assembly grammar is not logic; it uses the
methods of programming. Communication is mainly in command languages,
as in industrial orders, game rules and protocols, recipes, military com¬
mands, or governmental programs. The language is that of management, and
the sentences are like legal laws, not natural laws or valid truths. The basic
assumption in the action-oriented disciplines seems to be free will, while the
basic assumption in the empirical sciences seems to be determinism. I sub¬
mit that these are presenting laws of nature but belong to the metalanguage
of the discipline. They are decisions on how their statements are to be
viewed and organized in the language.
Each of the two extreme types of discipline has a few sentences belonging
to the other type. Management peppers its prescriptive commands with
descriptions of the conditions for alternative actions or initial states or the
goal to be achieved. Theoreticians pepper their extensive descriptions of
the states of their worlds with verification procedures (specified often in the
metalanguage they use in discussing the methods).
Scientists of the operational type demand a better balance. But it is the
disciplines called professions, whose mission is to transfer knowledge into
action, that employ the indicative and the imperative in more equal doses.
(But note that even medicine has its surgeons and internists; law, its prac¬
titioners and reviewers.)
I have elsewhere detailed the development of disciplines into phases that
mainly describe the complexity of the expressions in their languages from
the primitive noun system (taxonomic) to the sophisticated metalinguistic
expressions (formal deductive or simulative). [Gorn, 1965.] My claim has
been that an advanced stage in the development of a discipline’s language
and symbols makes it develop a contact with mathematics, and an even
more sophisticated—in fact, self-conscious—stage brings it into contact
with informatics. However, before I consider the ideology of informatics
that causes this relation to other disciplines, I should say something about
what I have called the mass ego of a discipline.
The claim that every discipline whose language is sufficiently sophis-
130 SAUL CORN
'I have called this property of some machines and their languages unstratified control. The
metalanguage of programming is contained in the object language. Hofstadter calls this property
“a strange loop.” [Hofstadter, 1979.]
INFORMATICS: IDEOLOGY, METHODOLOGY, SOCIOLOGY 131
now see how different people in the field were able to stress different kinds
of activity.
The formal theoretical aspects close to logic and mathematics cover such
topics as the theory of computation, the analysis of algorithms, switching
theory, automata theory [Hartmanis, 1979], the study of formal languages
[Greibach, 1979], the study of data structures, and the syntactic representa¬
tion of semantics. Among these topics, one learns what can and what cannot
be computed and for those problems that can be computed, what varieties
demand what kinds of computing rhythms and storage controls and at what
cost. One also learns the capabilities of various programming techniques,
the symbol-manipulative aspects of formal logic and theorem-proving, and
the formal techniques of program verification. [Ibid.]
The informatic activity that is empirical and experimental rather than
formal turns up in modeling human symbolic behavior in heuristic searching
and problem-solving. [See Newell and Simon, 1976a.] Whenever theory
informs us that some class of problem is unsolvable, we immediately search
for subclasses of problems that are solvable and do so by experimenting with
new solution methods until our theoreticians can prove to us that the
methods we find are obvious, limited, or fallacious. Even when theory in¬
forms us that a general-solution method exists but is very costly, we thrash
around for heuristic procedures that may fail to be universally successful
(these are called algorithms) but are likely to succeed much sooner. Much of
this modeling of human problem-solving behavior that remains empirical is
the subject of the area called artificial intelligence. [See also Amarel, 1976;
Newell and Simon, 1976a; Salton, 1976; Simon, 1976; and Pearson and
Slamecka, 1976.] This part of informatics is the part that is applied to the
area of engineering called robotics.
This empirical side of informatics, together with the theoretical side called
analysis of algorithms, is the area some call a major portion of informatics; it
is known as the design of algorithms. If we note that these procedures called
algorithms have equivalent representations in software and hardware, we
recognize our basic concept of processor again; this area is, therefore, what
I called the synthesis and analysis of interpreting processors. However, we
should not forget that in addition to the algorithms, there are devices that
feed them and collect their results, which are similar to human receptors and
effectors. These input and output devices can extend the semantics of the
symbol systems being processed to the various extensions of our senses and
motor controls needed in robotics. Thus, the various studies in graphics,
vision, and speech recognition, and their use in image processing and pattern
recognition, are also extensive portions of empirical informatics.
So far, we have examined the areas of informatics that are most scientific
in the recent English sense of the word. In both the theoretical and empirical
portions, people are concerned with truth—the truth about symbol manipu¬
lation. But many are engaged in informatic activity because of aesthetic or
ethical reactions. Some are attracted to the beauty, and some, especially
INFORMATICS: IDEOLOGY, METHODOLOGY, SOCIOLOGY 135
I remarked early in this paper that the word ideology has acquired a number
of meanings; I have been using two of those meanings. The first refers to a
systematic body of concepts, especially as applied to a manner or the con¬
tent of thinking characteristic of an individual, group, or culture; in our case,
the relevant groups or cultures are called disciplines, and the body of con¬
cepts includes what Thomas Kuhn calls paradigms. The second meaning is
that of Destutt de Tracy2 and the ideologues [Picavet, 1891 and 1971] around
the time of Napoleon. This meaning was the study of the development of
ideas from sensations, especially in humans; the ideologues intended ideol¬
ogy to be a branch of zoology, specifically of human biology, particularly
neurology and psychology. Destutt de Tracy’s [1817] book on ideology ac¬
knowledges descent from Condillac and has a second part on grammar and a
third part on logic. Clearly, he would have included today’s psycholinguists
and such developmental psychologists as Piaget among the ideologues. He
believed, for example, that ideas were initiated by direct perception and
extended by inner perceptions called memory, judgments (the comparison of
outer perceptions with memories), and will (the perception of desires and
action caused by this perception).
Our description of informatics and its relation to other disciplines, in fact,
of the generation of ideologies of disciplines and their Kuhnian transforma¬
tions, extends the basis in perception that the ideologues insisted on to a
basis in the extended perceivers and controllers beyond the human spec¬
trum, one that is provided by technology. In the case of informatics, it is the
electronic computer that provides this extension, just as it is the accelerator
in modern physics, the electron microscope in material science and biology,
and similar devices and techniques in genetics and anthropology and as it
was the microscope and telescope in the time of Galileo.
A true unification of human knowledge (such as the sciences) and action
(such as the arts and professions) would call for the ability to patch the
variety of ideologies together when they have been generated by extensions
in many directions beyond the human range. Such patchwork is going on in a
changing their scope, knowledge about each discipline will never be needed
in the same order or in exactly the same language in which it was produced.
Thus, the organization of disciplinary information as it is produced will
never be the same as the appropriate organization needed to recall it effi¬
ciently for future use. It follows that the organization of the library will
always need changing and, hence, the problems of library organization will
never be solved once and for all. Thus, the library profession will never be
static enough to be totally mechanizable. Each generation of librarians may
suffer a certain amount of technological obsolescence, but the profession as
a whole never can. Since its main problem is unsolvable, it is guaranteed to
be a stable profession.
In fact one can say, conversely, that only those professions will be stable
and, at long range, safe from technological obsolescence whose problems
are unsolvable. For example, physicians, whose main duty is to keep us
from dying, will always be with us. Similarly for teachers, whose job it is to
keep mankind from becoming stupid; or mass-communication professionals,
who must keep us informed.
It is attempts to solve new problems, especially the unsolvable ones, that
are responsible for establishing new disciplines and their ideologies, espe¬
cially the stable ones. The application of computers to these problems is,
therefore, likely to present symbolic simulations of their pragmatic de¬
mands, hence, of their ideologies. I would therefore expect a number of
professions to keep spinning off from informatics.
As a young discipline, however, theoretical aspects are still closely con¬
nected with the action-oriented design and handling of computational and
linguistic mechanisms. Fundamental results from the mathematical theory of
computation present a host of provably unsolvable problems of symbol ma¬
nipulation and a host of impracticable solutions to overly general problems
together with proofs that they cannot be improved. They mark the limits of
computation that every professional informatician must be aware of, as well
as the unsolvable areas that can be conquered only by subdivision. This
essential tie of the knowledge-oriented base to action-oriented aspects and
the consequent empirical activity may succeed in keeping informatics from
separating into distinct knowledge-oriented disciplines, the way mathemat¬
ics did after the Pythagoreans and Plato; grammar, after the Stoics; and
logic, after the logistic and formalist schools. This fact, coupled with the
host of unsolvable problems, may maintain informatics as a stable discipline
without its unravelling into separate sciences, arts, and professions. If such
is the case, informatics will not become a metascience, as did the just-
mentioned disciplines, or, for that matter, metaphysics.
In short, we should not try to separate computer science from information
science but should try to maintain one discipline of informatics. Any attempt
to cause such a separation by, for example, trying to create another metasci¬
ence [see Otten and Debons, 1970] would separate the action from the
140 SAUL CORN
3For further insights, see Furth [1974], Heilprin [1974], D. Hillman [1977], Kochen [1974],
Otten [1974], Sager [1977], Slamecka and Pearson [1977], Whittemore and Yovits [1974], and
Yovits, Rose, and Abilock [1977],
PERSPECTIVES ON INFORMATICS
AS A SEMIOTIC DISCIPLINE
The purpose of this paper is to analyze the logical and methodological rela¬
tions between information science and computer science and some of the
other subject areas that contribute to the scientific study of information.
Gorn proceeds toward this destination in two stages. First, he develops two
main theses: (1) Informatics (encompassing information and computer sci¬
ence, engineering, and technology) is a semiotic discipline; and (2) informa¬
tion science and information engineering should not and perhaps cannot be
separated very much. These theses provide a perspective that is then used to
analyze the relation between informatics and several neighboring disci¬
plines.
The gist of our comments may be summarized as follows. We are in full
agreement with the first thesis—that informatics is a semiotic discipline.
Given this perspective, we also fully concur that most of the neighboring
disciplines (including robotics, cybernetics, what is currently called informa¬
tion science, etc.) are professions or technologies rather than sciences and
that most uses of the word science in this area are honorific rather than
scientific. On the other hand, we are not convinced that information science
and informatic engineering should never be separated and that the honorific
use of the term science by information professions and technologies will
necessarily persist.
Our sympathetic regard for viewing informatics as a semiotic discipline is
easily explained: This view has been the common denominator of the aca¬
demic programs of Georgia Tech’s School of Information and Computer
Science since its establishment in 1964, and a major if not central thrust of its
research programs. [Slamecka and Gehl, 1978.] Thus, Gorn’s and our views
have largely coincided for nearly two decades [see Gorn, 1963; Slamecka,
1968]; and our research interests contributed to several of the topics men¬
tioned in his paper, notably the theorem of hardware-software equivalence
[Poore, Baralt-Torrijos, and Chiaraviglio, 1971], logic of plans and programs
[Baralt-Torrijos, 1973], and particularly empirical semiotics [Pearson,
1977]. Our comments on Gorn’s first thesis are thus largely an amplification
141
142 CHARLS PEARSON AND VLADIMIR SLAMECKA
'The word semiotics should not, however, be attributed to the American philosopher, Charles
Morris. Morris never called the discipline semiotics; he consistently called it semiotic, as did
also the founder of modern semiotics, the American logician and philosopher, Charles Peirce.
Modern usage of the term semiotic stems from its reintroduction into philosophy by the British
philosopher John Locke. The expression semiotics came only later, and Morris himself noted
that someone else had called the discipline semiotics. The anthropologist Margaret Mead is
often given credit for coining this term.
INFORMATICS AS A SEMIOTIC DISCIPLINE 143
2 It may be of interest in this regard to note that Morris was a student of George Herbert Mead,
who worked quite closely with John Dewey at the University of Chicago and that Dewey, if not
actually a student of Peirce’s, studied him quite thoroughly.
144 CHARLS PEARSON AND VLADIMIR SLAMECKA
The main issue of the paper, as far as the purposes of the entire study are
concerned, is the relation between informatics and some of the other disci¬
plines that contribute to the scientific study of information. Gorn draws a
distinction between the knowledge-oriented component of informatics (in-
formation-and-computer science) and its action-oriented component (infor¬
mation/computer engineering).
The inseparability of information science and computer science is well
worth emphasizing. Even though a few scholars, including Gorn, have
viewed this field in such a way for years, it is only recently that the similarity
or identity of goals and approaches has begun to surface (attested to, for
example, by the 1982 Symposium on Empirical Foundations of Information
and Software Science). Referring to our earlier comments, a note is appro¬
priate, however, regarding the scope of information and computer science.
Gorn refers to information-handling machines as symbol-manipulating am¬
plifiers, a definition that would seem to restrict such machines to digital
computers only—a restriction probably not intended. While symbols may
146 CHARLS PEARSON AND VLADIMIR SLAMECKA
play some role in many analog computers, it is the iconic structure that plays
the essential role; analog computers are icon-manipulating amplifiers. Simi¬
larly, simulators (in the engineering-training, not the software sense of the
term) are indexical information processors or index-manipulating amplifiers.
A key point of Gorn’s paper is that the knowledge-oriented and the ac¬
tion-oriented components of informatics cannot and should not be sepa¬
rated. The arguments supporting this belief need to be carefully assessed.
Informatics is perceived, in a fashion more characteristic of information
engineering, as being concerned with synthesis and analysis of symbolic
expressions and processors. The action-oriented (technological) and the
knowledge-oriented (scientific) aspects of computation are said to be less
likely to move apart, because the description of symbolic expressions re¬
mains tied to the prescriptive language of its interpreting processors, hence,
invariably forming a pragmatic tie.
Two comments seem appropriate here. First, the merger of the knowl¬
edge- and action-oriented components of informatics is a historic fact. The
mixture of knowledge and action does, indeed, maintain a pregnant environ¬
ment that can be harmed by forcing their separation; the late Mortimer
Taube, perhaps the first philosopher of informatics, was fond of saying that
there is no such thing as pure and applied science, only good and bad
science.
On the other hand, the merger of knowledge- and action-oriented compo¬
nents of informatics is not absolutely necessary. The pragmatic tie between
them refers to the pragmatics of doing technology, not to the pragmatics
studied by a triadic science. At times, such a separation may be necessary
for the major idealizations and abstractions characteristic of the great ad¬
vances in science to take place. To argue as Gorn does that such a separation
will of necessity “cause the cessation of the very effervescence the mixture
of knowledge and action maintains,” is to argue that the separation of
mechanical engineering and physical thermodynamics in the nineteenth cen¬
tury was also a similar mistake. While the original inventions of the steam
engine and the internal-combustion motor were made by practitioners, no
one would argue today against the drastic revolution in the level of technol¬
ogy enabled by the pure knowledge-engendered advances made in physical
understanding by such action-indifferent scientists as Gibbs, Maxwell,
Helmholtz, Kelvin, and others. This is effervescence at its best!
In informatics, such advances may be sought by standing back from the
interpreting process to look at the prescriptive language of interpreting pro¬
cessors, and then describing abstractly what it is they do. This, in fact, is one
reason for the necessity of introducing semiotic analysis into informatics—
to abstract from the study of pragmatic structure and to describe it indica-
tively.
Gorn is entirely correct in his assessment of other informatics-related
disciplines as being largely professions. The interesting movement we are
beginning to note is the emigration of areas from informatics to other profes-
INFORMATICS AS A SEMIOTIC DISCIPLINE 147
Alan J. Perlis
Vladimir Zwass
I am dividing my comments on Saul Gorn’s paper into three parts. In the first
part, I shall discuss the definition of computer science; in the second, I shall
reflect on the epistemological domain of this discipline; and in the third, I
shall assess the place of computer science among the family of disciplines
concerned with the study of information.
There remains no doubt that all fields of inquiry (and many other pursuits)
have been affected by computers. It is also evident that virtually all disci¬
plines have benefited from both computer science and information science
and that few, if any, have suffered undesirable dislocations due to their
influence.
COMPUTER SCIENCE AS THE
SCIENCE OF DISCRETE
MAN-MADE SYSTEMS
Joel Moses
I am pleased to have been given the opportunity to review Saul Gorn’s paper
on informatics and present some of my own views on computer science and
its relation to its neighboring disciplines. It is surprising to me that there
have been relatively few attempts to define such an important and fast¬
growing field as computer science and contrast its approaches with those of
mathematics, engineering, the physical sciences, and so forth.
We are well beyond the period (1930-1960) when people were fascinated by
the sheer fact that man-made artifacts could be general computing engines.
This fascination led to an overemphasis on the logical and dynamical aspects
of information processing. The fact that computers actually performed com¬
putations by manipulating numbers and other symbols, led some theoreti¬
cians to studies of the logical limits of various computing machines (e.g.,
finite-state machines, Turing machines), and others to philosophical explora¬
tion of the meaning of certain computational devices such as assignments of
values to variables. In the last two decades, the emphasis on the dynamics of
computation led to studies of how fast one can perform certain computations
such as multiplication or to what extent one can trade space for time in a
variety of algorithms.
While these studies have been interesting, their relevance to the practice
of computing has been limited except in special areas. I view the rise of
software engineering, led largely by Edsger Dijkstra, as a reaction to much
of the research on computational dynamics. The position I take here is
largely a systems view of computer science that is characteristic of the
undergraduate program in computer science at the Massachusetts Institute
of Technology. In this approach, large software systems are seen as rela-
157
158 JOEL MOSES
tively static objects that are worthy of serious study. Experience shows that
the cost of designing, modifying, and maintaining large systems grows non-
linearly with the size of the system. Furthermore, large systems that one
wishes to remain viable must be changed constantly as the environment in
which they operate changes. Methods for coping with the complexity that is
often present in large systems, and, hence, for easing the process of embed¬
ding changes in specifications of these systems, are among the key issues in
static analyses of systems.
Although I emphasize here the static aspects of computation, it is not my
intent to imply that other aspects, such as the dynamic or aesthetic, of the
field should be entirely ignored. Efficiency, or dynamic issues underlying a
given system, should be dealt with after the static structure has been de¬
fined. Making it easier for someone to use man-made systems involves a
variety of what are often aesthetic issues. Such issues are dealt with by
disciplines other than computer science, such as art, linguistics, and psy¬
chology. Computer scientists have increasingly been getting involved in
such work, often in collaboration with those in the other disciplines.
A view of computer science that emphasizes the static rather than the
dynamic aspects of computation yields distinctions different from those
Gorn proposed between computer science and its neighboring disciplines.
Computer science is clearly not a physical science. The physical sciences
are concerned with discovering the principles of the design of a single sys¬
tem, namely God’s. Computer science deals with principles for creating
new, large man-made systems and is not limited to studying computer sys¬
tems. God’s system is indeed complex, but new approaches for reducing
complexity that are absent in God’s original design are not of much interest
to physical scientists.
Software engineers who are, in the main, very close to the systems view
espoused here, seem to claim an affinity between computer science and
engineering. Clearly, they have a point, but the analogy ought not to be
stretched too far. Engineers certainly try to create new man-made systems;
they are, however, limited by the properties of the physical world. They
spend most of their time figuring out ways to circumvent limitations in
materials or mechanical and electrical devices. Computer scientists are in
the enviable position of being given primitives (i.e., instruction sets) that can
usually be considered to work perfectly, without the interference of the
demons that God has placed in the primitives with which engineers must
cope. If computer science is to be viewed as a branch of engineering, it
should be considered as abstract engineering, largely unconstrained by limi¬
tations of the physical world.
Physical scientists and engineers are making increasing use of computa¬
tion in their work. The reason for this is that computers can make highly
SCIENCE OF DISCRETE MAN-MADE SYSTEMS 159
PURE MATHEMATICS
There is one field that is as abstract, in fact, more abstract than the picture I
have painted of computer science; that field is pure mathematics. Mathemat¬
ics missed out on being the science of man-made systems for several rea¬
sons. The four main areas of pure mathematics are algebra, analysis, geome¬
try, and topology; the last three of these all deal with continuous rather than
discrete models. Abstract algebra, to which I shall return later, has been
used as a tool in the other areas and usually not as a model of discrete
systems. The three areas in mathematics (i.e., elementary number theory,
combinatorics, and logic) that have been applied to discrete systems have
not been at the core of modern mathematics. These areas have, in fact, been
heavily used in the dynamical and logical models of computer science.
The second failing of modern mathematics was its general lack of interest
in issues of efficiency that are at the core of the dynamical view of computa¬
tion and of great importance in man-made systems. Thus, investigators in
computer science and operations research have had to develop new mathe¬
matics for discrete systems, with relatively little support from mathemati¬
cians.
160 JOEL MOSES
I now proceed to discuss in general terms what I perceive are some of the
core issues arising in large, discrete man-made systems. The principle issue
is the control of complexity. Something is complicated, the dictionary says,
when it has a large number of interconnected parts. Thus, complexity is a
function of the static nature of a system. It is unfortunate that theoreticians
in computer science use the term complexity theory for the theory that deals
with such issues as the minimum time or space needed to solve specific
problems. Such a dynamic theory ought to have been called a theory of
difficulty, but, one supposes, such a term is not as exciting as complexity
theory. In any case, apparently difficult problems, such as factorization of
integers into primes, need not be complex and vice versa.
If complexity is measured by the number of interconnections, then the
obvious way to achieve simplicity is to reduce the number of interconnec¬
tions a system requires. There is a limit to such a reduction when the number
of components (procedures, people, transistors, musical notes, etc.) has to
be large in order to achieve the desired functionality. The real key to reduc¬
ing complexity in large systems is to introduce additional structure or regu¬
larity when this is possible. When the interconnections possess some struc¬
ture, one is capable of modeling the systems using abstractions. In a more
abstract setting, the number of components and their interconnections is
much reduced. For example, if one recognizes that the same process is being
performed in many parts of the system, one can introduce a single new
procedure. Similarly, if the same action is being performed on a number of
elements, one can organize the elements into a vector. In the first case, we
have a procedural abstraction, in the second, a data abstraction.
An instance of an organizational structure of relatively low overall com¬
plexity is that of a pyramid or tower of linguistic abstractions. For example,
computer systems are often organized as layers of languages: a language
(actually, a family with restricted interconnections) of electronic compo¬
nents, a microprogramming language, an assembly language, and a higher
level language. Application programs can also often be designed in such a
pyramidal fashion; unfortunately, this is not always possible. For example,
sorting-routines are in general too combinatorial in nature to permit such
restructuring. Nevertheless, I believe it would be wise to consider such
designs as objectives for large systems.
SCIENCE OF DISCRETE MAN-MADE SYSTEMS 161
As can be seen from the preceding discussion, my view is that the science
underlying computer science has little to do with computers. Rather, it is the
systematic study of issues related to the design of discrete man-made sys¬
tems. Small systems are often of little interest in such studies, since one has
far less difficulty in designing them, except when circumventing the vagaries
of nature or in producing some pleasing effect.
162 JOEL MOSES
SYMBOLIC MANIPULATION
Peter Wegner
163
164 PETER WEGNER
In his paper, Gorn examines the roles of ideology, methodology, and sociol¬
ogy in the development of disciplines. He points out the close relation be¬
tween ideologies (systematic bodies of concepts) and paradigms. He dis¬
tinguishes between knowledge-oriented activities such as mathematics,
action-oriented activities such as farming, and activities that transform
knowledge into action, such as education and engineering. He suggests that
sciences cannot flower independently of technology and in particular that in¬
formatics (computer-and-information science) could not begin to flower be¬
fore the technology of computing was developed. Informatics was initially
action-oriented (practical) and has achieved a fusion of action-oriented and
knowledge-oriented activities as it has matured. This balance is one of the
strengths of computer science in that practical results suggest relevant
theory, and theoretical results can increase the effectiveness of action. Gorn
warns that a divorce of action-oriented from knowledge-oriented activities
could lead to disaster.
Gorn’s choice of ideology, methodology, and sociology as characterizing
attributes of disciplines is unconventional but suggestive. These attributes
are independent of particular disciplines and are applicable in characterizing
any discipline or human activity. The ideology of a discipline determines its
paradigms and its criteria for evaluating the quality and relevance of contri¬
butions to the discipline. Its methodology is the principles that underlie the
tools and techniques used in pursuing the goals determined by an ideology.
Its sociology includes social interactions among practitioners of the disci¬
pline and the social impact of its concepts and products on society. Thus, the
ideology, methodology, and sociology associated with a discipline may be
thought of as its paradigms, underlying principles, and impact on people.
Ideologies are associated with young people and young disciplines. Com¬
puter science became an academic discipline in the mid-1960s with the emer¬
gence of the first departments of computer science and the publication of a
comprehensive undergraduate curriculum. [ACM Curriculum Committee,
1968.] The search for an identity led to a number of alternative definitions of
computer science, associated with different paradigms forjudging the value
of research. In Gorn’s terminology, these definitions express alternative
ideologies.
The first definition reflects an empirical view, since it asserts that com¬
puter science is concerned with the study of classes of such man-made
phenomena as computers, programming languages, algorithms, and data
structures. The second and third definitions reflect a mathematical tradition,
since algorithms and information structures are two abstractions that deter¬
mine different paradigms for modeling the phenomena of computer science.
The fourth definition takes the viewpoint of engineering and reflects the
great complexity of problems of information engineering encountered in
developing complex hardware-software systems.
Each of these definitions has motivated valuable contributions, and they
should be regarded as complementary rather than mutually exclusive. The
first was dominant in the 1950s and is experiencing a resurgence with the
increasing fashionability of empirical computer science. [Feldman and
Sutherland, 1979.] The second led to the flowering of research in the areas of
analysis of algorithms and computational complexity in the late sixties and
early seventies. [Arden, 1980a.] The third has motivated the study of seman¬
tic models of programming languages, also in the late sixties and early
seventies. The fourth has led to software engineering, which emerged as a
subdiscipline in the late 1960s and became a dominant research area in the
mid-1970s.
The first view of computer science is action-oriented, the second and third
are knowledge-oriented, and the fourth (software engineering) represents a
synthesis of both approaches. Thus, the dominant paradigm was action-
oriented in the 1950s, knowledge-oriented in the late 1960s and early 1970s,
and evolved as a fusion of the two in the late 1970s and 1980s. Computer
science was action-oriented in the period immediately following the birth of
computer technology, as there were no theories to support the new technol¬
ogy. The development of a theoretical foundation led to a brief period of
dominance of theoretical approaches, followed by the current period in
which the ideal, not always realized, is research that combines theory and
practice.
These four definitions span a variety of views of computer science. We
must ask whether this variety is a manifestation of the confusion of a young
discipline or whether a variety of paradigms is an inherent (permanent)
feature of computer science. My view is that variety is both inherent and
intrinsically healthy (many flowers should continue to bloom). The informa¬
tion sciences will blossom into a broad spectrum of mathematical, experi¬
mental, and engineering subdisciplines. By the year 2000, many universities
may have a school of information sciences with several departments sharing
a common set of core requirements, just as schools of engineering currently
include departments of electrical, mechanical, and civil engineering.
Schools of information sciences will contain information engineering as a
major component; information engineering will, in turn, contain software
engineering and knowledge engineering as two subdisciplines. Software en¬
gineering is concerned with managing the complexity of software, while
166 PETER WEGNER
During the period 1950-1980, there have been radical changes in the eco¬
nomics of information processing. The cost per executed instruction of com¬
puter hardware has decreased by a factor of two every two or three years,
while the complexity and total cost of computer applications has increased
by several orders of magnitude. The cost of software as a proportion of total
computing cost has increased from under 20 per cent in 1960 to over 80 per
cent in 1980.
Programming languages such as FORTRAN, originally developed for nu¬
merical problems with several hundred lines of code, proved to be inade¬
quate for large applications, with real-time requirements and hundreds of
thousands of lines of code, concerned with controlling the operation of
ships, airplanes, banks, or chemical plants. Large software projects in the
1960s and 1970s frequently failed to meet schedules, were subject to enor¬
mous cost-overruns, and sometimes had to be abandoned because their
complexity became unmanageable.
In the late 1960s, it was recognized that there was a software crisis and
progress in software technology required the development of systematic
techniques for managing the complexity of large software systems. [Naur,
Randell, and Buxton, 1976.] The discipline of software engineering was born
with the aim of providing a technological foundation for the development of
software products analogous to that provided by conventional engineering
for the development of physical products.
Software engineering exemplifies the fusion of action-oriented and knowl¬
edge-oriented activities. It is motivated by the economic considerations of
making software cheaper and more reliable. This contrasts with the schol¬
arly motivations traditionally associated with empirical and mathematical
paradigms. However, economic motivations are not necessarily inferior to
scholarly motivations as a basis for research. On the contrary, economically
PARADIGMS OF INFORMATION ENGINEERING 167
part of its users. It is sponsoring not only language design and implementa¬
tion but also the study of methodology, education, and technology transfer.
For example, I am developing guidelines for education and technology trans¬
fer that will include recommendations on how to teach novel language fea¬
tures; structure ADA courses for managers, programmers, and other con¬
stituencies; and introduce ADA into industrial, military, and university
programming environments. [Wegner, 1982.] This task is difficult since there
are no guidelines for developing such guidelines. (Who will guide the
guides?) My approach is to provide information about current education and
technology-transfer efforts and draw attention to issues without necessarily
resolving them.
ADA has, within the short period of five years, generated a subculture of
tens of thousands of devoted adherents throughout the world. Hundreds of
companies are involved in a variety of ADA-related activities ranging from
implementation to education. The rapid spread of ADA at a time when it is
not yet implemented and its methodology is not yet tested illustrates Thomas
Kuhn’s assertion that inadequacy of a current paradigm provides fertile
ground for new paradigms.
become one of the largest employers of labor and of capital. The nature of
computer applications is changing from primarily numerical applications,
which compute a numerical result, to data management and embedded com¬
puter applications, whose objective is to control a larger system whose
purpose is not primarily computational. Computer control is playing a larger
and increasingly important role in the management of society. Power is
shifting from the industrial establishment to those who control the dissemi¬
nation and distribution of information.
Computers are playing an increasing role not only in the management of
society but also in the management of knowledge. After many false starts,
artificial intelligence is finally reaching a level that allows building expert
systems whose performance is competitive with human experts. Personal
computers are becoming cheaper and more powerful, to the point where
everyone soon will be able to afford a pocket computer that executes a
million instructions a second, has a large memory bank extensible with
cassettes, and supports graphical and voice communication. The combina¬
tion of greater power and greater accessibility will entirely change the rela¬
tion between man and computers, so that computers will no longer be
merely problem-solving tools. They will become an extension of man’s intel¬
lect and serve to amplify intellectual capability in the same sense that con¬
ventional machines amplify man’s physical ability. The new role of com¬
puters will give rise to new paradigms for managing and using knowledge
and to a new subdiscipline of computer science, which I will refer to as
knowledge engineering.
KNOWLEDGE ENGINEERING
users, including multiple views and other forms of redundancy, rather than
efficiency and precision for the benefit of computers. Whereas knowledge
structures for computer understanding must be very detailed and precise,
knowledge structures for human understanding are concerned, not with the
precise specification of a computational task, but with organizing knowledge
for human readers who possess considerable contextual understanding and
are capable of conceptualizing at a level far above that of the computer.
Restructuring existing knowledge so that it is more accessible to humans
involves more than putting existing knowledge repositories such as the
Library of Congress on computers and accessing them through information-
retrieval systems. It involves restructuring existing knowledge so that it can
be flexibly presented in different formats for different contexts of use. The
technology for such restructuring is not well understood, but its nature can
be illustrated by considering recent developments in computerized printing
technology and computer-based learning.
Computers are revolutionizing printing technology and allow high-quality
text to be quickly and cheaply produced. Word-processing systems provide
authors with much greater control over production, layout, and modification
of text. Soon computers will be used not only for writing and printing books
but also for reading them. Book-size computers with flat panel displays will
make electronic books a reality. The greater bandwidth of man-machine
interfaces will qualitatively change the nature of man-machine communica¬
tion and make communication of knowledge by reading computer books
more effective than conventional communication by reading hard-copy
books.
Whereas hard-copy books consist of a linear sequence of pages, materi¬
als intended to be read on a computer may have a graph structure with
different entry points for readers with different backgrounds. Multiple win¬
dows allow the reader to pursue several lines of thought simultaneously or
view a given object at several levels of detail. Interactive responses by the
user can be used by the computer to tailor the mode of graph traversal to the
interests and skill level of the student. Each mode of the graph structure can
include dynamically animated pictures, texts, and programs. For example,
the mathematician may wish to animate the development of a proof, while
the computer scientist may wish to animate the process of program develop¬
ment and execution.
An electronic book represents a family of different hard-copy books that
could be obtained by printing out nodes of the graph structure in a particular
linear order for particular kinds of students. It is conjectured that flexibility
in adapting the pace and order of presentation of information to the student,
combined with the power of animation (possibly augmented by voice input
and output) can, if properly used, enormously increase the student’s capac¬
ity to absorb and understand both elementary and advanced knowledge.
Knowledge graphs that may be entered at different points and traversed in
different ways represent a paradigm for knowledge engineering that imposes
PARADIGMS OF INFORMATION ENGINEERING 173
COMPUTER-BASED LEARNING
text on the same subject matter but will require authors to organize the
material in new ways. The modular approach has the disadvantage that it
may violate the natural continuity of the subject matter but the advantage
that it requires authors systematically to decompose knowledge into man¬
ageable modules.
Creators and readers of a computer textbook form a social community
whose members can communicate directly with each other via a computer
message system. Authors can incrementally make the text available, receive
instantaneous feedback from readers, and rapidly respond to such feedback.
Man-machine communication may be used not only for machine display of
knowledge, but also for communication among its community of creators
and users. Such social interaction will permeate all work in knowledge en¬
gineering and may fundamentally affect the sociology of all disciplines by
providing a new mode of communication among scholars.
Computers will not only help students to learn more effectively, but also
help authors to write more effectively. Conventional printing technology
permits only very costly enhancement of a book by means of new editions.
Computer technology permits continuous incremental enhancement after
development of the book has been completed, thereby permitting im¬
provements in quality and flexible adaptation to changing requirements that
would previously have been impossible.
The advantages of incremental enhancement can be illustrated by drawing
an analogy between the life cycle of a program and a book. Studies have
shown that 80 per cent of the effort of supporting a program over its life
cycle is in maintenance and enhancement. With conventional printing tech¬
nology, the only form of maintenance and enhancement is printing a second
edition, which is time-consuming and expensive. By allowing cheap incre¬
mental maintenance and enhancement, computer printing technology could
completely change the role of authors in the life cycle, allowing them to play
a much more active role in both the production and enhancement process.
A computer support system for a community of authors and students
involved in creating, disseminating, and learning a body of knowledge will
be called an educational environment. An educational environment should
support not only individual students but also communication among the
community of students and authors. An educational environment should
provide interactive feedback to students using the system and interactive
feedback to authors concerning the effectiveness of the system. It should be
concerned not only with technical issues of knowledge representation but
also with social issues of communication among a community of users with
different goals. As the man-computer interface becomes more powerful and
computers are increasingly used for communication among humans, social
issues and human factors for effective communication will become increas¬
ingly important.
An educational environment must provide an author environment that
helps authors create and enhance text modules, a student environment that
PARADIGMS OF INFORMATION ENGINEERING 175
CONCLUSION
Saul Gorn
In the spectrum of comments on my paper, the one that was furthest re¬
moved from my point of view was that of Joel Moses. He remarks on the
irony of his individual interest being symbolic manipulation when he cannot
accept my making symbolic manipulation a central issue of computer sci¬
ence.
I gather that he wants to identify the field as being concerned with the
analysis, design, and management of large, complex, discrete, man-made
systems. He would, therefore, expect it to include certain chunks of cyber¬
netics, especially of systems science, systems engineering, decision science,
and management science. The involvement of computers, on the one hand,
and symbol manipulation, on the other, may be absolutely essential in han¬
dling problems but are not, from his point of view, the central issue. To him,
a part of what I have called informatics that is more characteristic of the area
he is discussing would be software engineering. Peter Wegner in his com¬
ments also seems to consider this to be the most characteristic area of
informatics; I will comment on this later.
Now, whether the area described by Moses will actually crystallize into
what I have called a discipline will, of course, depend on such social phe¬
nomena as the development of a single new ideology and a few paradigms. In
his discussion of what could be considered an outline of such an ideology, I
found a few of his attitudes a bit puzzling. First of all, what led him to think
that I called the arts and crafts knowledge-oriented disciplines? Secondly,
Moses claims that it is the static structure of the large, discrete, man-made
systems with which he is concerned, when it is their maintenance and
robustness and modifiability in a constantly changing environment that is on
his mind. I can only assume that Moses believes it is the stability of the
system that is the central issue. Finally, and most important, I am puzzled
by his claim to be pragmatic while making a sharp distinction between en¬
gineering for God’s system (i.e., the physical world) and engineering for
man-made systems. Moses’s assumption seems to be that there is only one
world out there, with only one proper perception of it—a point of view I
would call monoideologicab, I think such an assumption is held by many
materialists, idealists, positivists, and “unified-scientists” and results in
rather complicated methodological positions.
177
178 SAUL GORN
JIn this connection, it is interesting to go back to the Proceedings of the Western Joint Com¬
puter Conference 1961, whose theme was Extending Man’s Intellect and look at the papers by
Herbert Simon, Edward Feigenbaum, and Julian Feldman. [Western Joint Computer Confer¬
ence, 1961.]
180 SAUL CORN
ever, it was my intention to emphasize that each discipline has its own flavor
of reasoning and these flavors may be a major concern of methodology; 1
have mentioned this in connection with expert systems.
Pearson and Slamecka are also correct about my not intending to restrict
attention to the digital in symbol manipulation. I have always felt that analog
machines should also be involved. They are better suited and faster in imitat¬
ing by feedback the adaptive and conditioning type of learning, because they
do not proliferate memories to be retrieved. Robotic sensors and receptors
can obviously use them.
When Pearson and Slamecka object to my saying that the symbology of
mathematics and physics may be formal, they forget, on the one hand, that
the intuitionists in mathematics do not accept formalist views and, on the
other, that rational mechanics is considered part of physics in the United
States but part of mathematics in Europe.
Pearson and Slamecka find that harnessing knowledge and action is re¬
lated to pragmatics merely in “doing technology, not to the pragmatics
studied by a triadic science.” But does not the action force the selection of
such using and interpreting behavior as is relevant to the symbology? In this
connection, I was surprised to see Helmholtz listed as an action-indifferent
scientist. His study of the adaptation time required in using prismatic lenses
was a direct study of the pragmatic effects of perception on action and a
model for psychologists.
We are agreed, however, that the professional aspects of informatics go
well with the theoretical, though we may argue about how separate they are
and for how long. Not only do I insist, as Pearson and Slamecka do, that
pragmatic questions should be an object of study within the field, but I
submit that professional aspects must also be included in such studies. It is
very much to the point, for example, that a contextfree language cannot be
interpreted by a finite-state machine but also needs a pushdown store. To
me, this is not merely a theorem in automata theory and formal languages
but also a fact of mechanical pragmatics; for it clearly states a relation
between a set of symbolic expressions and its mechanical user or inter¬
preter.
As for the possibility of a number of professional aspects having their
pilot studies within informatics and then graduating into separate disciplines
with related ideologies but different paradigms, I find Pearson and Slamec-
ka’s description and examples very well put. Except for the VLSI (very-
large-scale integration) example, they belong to the area of knowledge¬
engineering discussed by Wegner.
Wegner remarks early in his comments that knowledge-oriented, action-
oriented, and professional aspects of informatics have different paradigms;
and I have to agree that such radically different orientations must differ in
their paradigms. I believe, however, that a shared ideology is enough to keep
them in a single community. I therefore disagree with Thomas Kuhn that the
mark of a single discipline (physical science, in his case) is a single paradigm;
A PRAGMATIST REPLIES 183
Allen Newell
I thank Elaine Kant and Stu Card for comments on an earlier draft and Paul Birkel and Marc
Donner for leading me to write the paper. Note: This research was sponsored in part by the
Defense Advanced Research Projects Agency (DOD), ARPA Order No. 3597, monitored by the
Air Force Avionics Laboratory under Contract F33615-78-C-1551. The views and conclusions
contained in the paper are those of the author and should not be interpreted as representing the
official policies, either expressed or implied, of the Defense Advanced Research Projects
Agency or the United States government.
187
188 ALLEN NEWELL
one task after another, initially with programs that can accomplish them
crudely, followed gradually by successive refinements. Game-playing,
theorem-proving, medical diagnosis—each provides a single developmental
strand that can be tracked. Thus, a history of AI as a whole could be written
in terms of the geography of tasks successfully performed by AI systems.
Almost orthogonal to this task-dimension is that of the intellectual func¬
tions necessary for an intelligent system—representation, problem-solving
methods, recognition, knowledge acquisition, and so forth—what can be
termed the physiology of intelligent systems. All these functions are required
in any intellectual endeavor of sufficient scope, though they can be realized
in vastly different ways (i.e., by different anatomies), and tasks can be found
that highlight a single function, especially for purposes of analysis. Thus, a
history can also be written that follows the path of increased understanding
of each function and how to mechanize it. Both of these structural features
of AI, and perhaps especially their matrix, provide potentially fruitful
frameworks for a history. Their drawback is just the opposite from the ones
mentioned earlier, namely, they lead to histories that are almost entirely
internal, shedding little light on connections between AI and neighboring
disciplines.
I settle on another choice, which I will call intellectual issues. It is a
sociological fact of life that community endeavors seem to polarize around
issues—fluoridation versus ban fluoridation, liberal versus conservative.
Such polarizing issues are not limited to the purely political and social arena
but characterize scientific endeavors as well—heliocentrism versus geocen¬
trism, nature versus nurture. Intellectual issues are usually posed as
dichotomies, though occasionally three or more positions manage to hold
the stage, as in the tussle between capitalism, socialism, and communism.
Intellectual issues are to be distinguished from issues in the real world of
action. No matter how complex and ramifying the issues of individual free¬
dom and state control that lie behind a fluoridation campaign, the passage or
defeat of an ordinance banning fluoridation is a concrete act and is properly
dichotomous. But with nature versus nurture, the dichotomy is all in the eye
of the beholder, and the real situation is much more complex (as is pointed
out ad nauseum). The tendency to polarization arises from the way people
prefer to formulate intellectual issues.
Scientifically, intellectual issues have a dubious status at best. This is true
even when they do not have all the emotional overtones of the previous
examples. Almost always, they are defined only vaguely, and their clarity
seldom improves with time and discussion. Thus, they are often an annoy¬
ance to scientists just because of their sloganeering character. Some time
ago, in a conference commentary entitled You Can’t Play Twenty Questions
with Nature and Win, I myself complained of the tendency of cognitive
psychology to use dichotomies as substitutes for theories (e.g., serial ver¬
sus parallel processing, single-trial versus continuous learning). [Newell,
1973 b.]
190 ALLEN NEWELL
I will actually do the following: I will identify, out of my own experience and
acquaintance with the field, all of the intellectual issues that I believe have
had some prominence at one time or another. Although I will take the field of
AI as having its official start in the mid-1950s, the relevant intellectual issues
extend back much earlier. We surely need to know what issues were extant
at its birth. I will attempt to put a date both on the start of an issue and on its
termination. Both dates will be highly approximate, if not downright specu¬
lative. However, bounding the issues in time is important; some issues have
definitely gone away and some have come and gone more than once, though
transformed each time. I will also discuss some of the major features of the
scientific scene that are associated with a given issue. I will often talk as if an
issue caused this or that. This is in general illegitimate. At best, an issue is a
publicly available indicator of a complex of varying beliefs in many scientists
that have led to some result. Still, the attribution of causation is too conven¬
ient a linguistic practice to forego.
Table 1 lays out the entire list of intellectual issues. In addition to the
short title ol the issue, expressed as a dichotomy, there is an indication of an
important consequence, although this latter statement is necessarily much
abbreviated. The issues are ordered vertically by date of birth and within
that by what makes historical sense. All those born at the same time are
indented together, so time also moves from left to right across the figure;
except that all the issues on hand when AI begins in 1955 are blocked
togethei at the top. Issues that show up more than once are multiply repre-
Table 1. The Intellectual Issues of Al
191
192 ALLEN NEWELL
sented in the table, according to the date of rebirth, and labeled #1, #2, and
so forth. When the ending date is not shown (as in Reason versus Emotion
and Feeling #7: 1870- ), then the issue still continues into the present.
The issues are discussed in historical order, that is, according to their
order in the table. This has the advantage of putting together all those issues
that were animating a given period. It has the disadvantage of mixing up lots
of different concepts. However, since one of the outcomes of this exercise is
to reveal that many different conceptual issues coexisted at any one time, it
seems better to retain the purely historical order.
We can start with the issue of whether mechanisms were essentially without
purpose. This is of course the Cartesian split between mind and matter, so
we can take Descartes as the starting point. It is an issue that can not be
defined until the notion of mechanism is established. It is and remains a
central issue for AI, for the background of disbelief in AI rests precisely with
this issue. Nevertheless, I place the ending of the issue with the emergence
of cybernetics in the late 1940s. If a specific event is needed, it is the paper
by Rosenblueth, Wiener, and Bigelow, which puts forth the cybernetic
thesis that purpose could be formed in machines by feedback. [Rosenblueth,
Wiener, and Bigelow, 1943.] The instant rise to prominence of cybernetics
occurred because of the universal perception of the importance of this
thesis. (However, the later demise of cybernetics in the United States had
nothing whatsoever to do with any change of opinion on this issue.) AI has
added the weight of numbers and variety to the evidence, but it has not
provided any qualitatively different argument. In fact, from the beginning,
the issue has never been unsettled within AI as a field. This is why I charac¬
terize the issue as vanishing with cybernetics. It does remain a live issue, of
course, in the wider intellectual world, both scientific and nonscientific,
including many segments of cognitive science. Above all, this issue keeps AI
in perpetual confrontation with its environment.
Intelligence presupposes purpose, since the only way to demonstrate in¬
telligence is by accomplishing tasks of increasing difficulty. But the relation
is more complex the other way around. While purpose could hardly be
detected in a device with no intelligence, that is, with no ability at all to link
means to ends, no implication follows about the upper reaches of intelli¬
gence. Animals, for instance, are obviously purposive yet exhibit strong
limits on theii intelligence. Thus, settling the question of artificial purpose
does not settle the question of artificial intelligence. The continuation of this
basic controversy throughout the entire history of AI over whether intelli¬
gence can be exhibited by machines confirms this separation. Yet, histori¬
cally it is not right to posit a separate issue of mechanism versus intelligence
to contrast with mechanism versus teleology. No such distinction ever sur¬
faced. Instead, there is an underlying concern about the aspects of mentality
INTELLECTUAL ISSUES IN THE HISTORY OF Al 193
that can be exhibited by machines. This shows itself at each historical mo¬
ment by denying to machines those mental abilities that seem problematic at
the time. Thus, the argument moves from purpose in the 1940s to intelli¬
gence in the 1950s. With the initial progress primarily in problem-solving, we
occasionally heard in the 1960s statements that machines might solve prob¬
lems but they could never really learn. Thus, the basic issue simply endures,
undergoing continuous transformation.
A critical issue for AI that had come and gone long before AI really began is
the issue of vitalism—do living things constitute a special category of en¬
tities in the world, inherently distinct from inanimate physical objects. As
long as this issue was unsettled, the question of whether the mind of man
was mechanical (i.e., nonspecial) was moot. It is difficult to conceive of
concluding that the animate world does not generally obey the laws of the
physical world but that the mind is an exception and is entirely mechanical.
Thus, only if vitalism has been laid to rest for our bodies can the issue be
joined about our minds.
The vitalist controversy has a long and well-chronicled history. Retro¬
spectively, it appears as an inexorable, losing battle to find something spe¬
cial about the living, though the issue was joined again and again. Organic
matter was just a different kind of matter from inorganic matter—an issue
laid to rest finally with the synthesis of urea, an indisputably organic mate¬
rial, from inorganic components in 1828 by Wohler. Organisms had their
own inherent internal heat—an issue laid to rest in the work of Bernard by
the mid-1800s. For our purposes, the starting and ending dates of the issue
are not critical. Vitalism’s last champion may be taken to be the embryolo¬
gist Hans Driesch at the turn of the century, who proposed that organisms
develop only by virtue of nonmaterial vital principles, called entelechies.
[Driesch, 1914.] Issues almost never die, of course, as the continued exist¬
ence of the Flat Earth Society should remind us. Nevertheless, no substan¬
tial intellectual energy has been focused on vitalism in more than fifty years.
That the human body is a physical machine, operating according to under¬
stood physical laws and mechanisms, sets the stage for considering the
mechanistic nature of thought and intelligence.
The basic separation of the heart from the head occurred long ago and is a
fundamental part of Christian folk psychology. It is background. What con¬
cerns us is the ascription of reason (cold logic) to machines and the belief
that a machine could have no heart—no feelings or emotions—to ever con¬
flict with its reason. I do not seem to find any good way to fix the initiation of
this issue. The striking characteristic of the golem of Rabbi Loew in 1580
194 ALLEN NEWELL
For science as a whole, the separation from philosophy and the acceptance
of empiricism as a fundamental tenet occurred centuries ago. For psychol¬
ogy, this occurred very recently, in the last decades of the nineteenth cen¬
tury. Indeed, psychology celebrates the establishment of the first experi¬
mental laboratory (Wundt’s in Leipzig) in 1879. It was not an especially
difficult passage for psychology, given the rest of science as a model. It can
be considered complete by the rise of behaviorism, say, by Watson’s classic
paper. [Watson, 1913.] Thus, this issue emerged and vanished before AI
began. The residue was a continuing tradition in philosophy concerned with
mind, which was completely distinct from work in psychology and, even
more so, from technology. This issue ensured that when AI did emerge,
which happened instantly on computers becoming sufficiently powerful,1 it
would be without more than peripheral involvement of the philosophy of
mind.
' A case can be made that serious AI started as soon as computers attained 4K of random-access
primary memory.
INTELLECTUAL ISSUES IN THE HISTORY OF Al 195
When computers were first developed in the 1940s, they were divided into
two large families. Analog computers represented quantities by continuous
physical variables, such as current or voltage; they were fast, operated
simultaneously, and had inherently limited accuracy. Digital computers rep¬
resented quantities by discrete states; they were slow, operated serially, and
had inherently unlimited accuracy. There was a certain amount of skirmish¬
ing about which type of computer was better for which type of job. But the
technical opinion-leaders maintained a view of parity between the two
196 ALLEN NEWELL
families—each for its own proper niche. Inevitably, there arose hybrid com¬
puters, which claimed to have the best of both worlds: digital control and
memory coupled with analog speed and convenience.
It was all over by 1970. The field of computers came to mean exclusively
digital computers. Analog systems faded to become a small subpart of elec¬
trical engineering. The finish was spelled not just by the increased speed and
cost-efficiency of digital systems, but by the discovery of the Fast Fourier
Transform, which created the field of digital signal processing and thus
penetrated the major bastion of analog computation. The transformation of
the field is so complete that many young computer scientists hardly know
what analog computers are.
The main significance of this issue, with its resolution, was to help create
the discipline of computer science and separate it from electrical engineer¬
ing. Its effect on AI lies mostly in the loss of an analytical point of view, in
which the contrast between analog and digital computation is taken as the
starting point for asking what sort of information-processing the nervous
system does. An admirable example of this point of view can be seen in the
notes for von Neumann’s Silliman Lectures, published posthumously, [von
Neumann, 1958.] This style of analysis belongs to the world of cybernetics
and not to that of AI. I doubt if many young AI scientists have read von
Neumann’s little book, though it was highly regarded at the time, and von
Neumann was one of the towering intellects of the computer field.
We now come to the first of the issues that characterizes AI itself, as op¬
posed to the background against which it emerged. The digital-computer
field defined computers as machines that manipulated numbers. The great
thing was, its adherents said, that everything could be encoded into num¬
bers, even instructions. In contrast, scientists in AI saw computers as ma¬
chines that manipulated symbols. The great thing was, they said, that every¬
thing could be encoded into symbols, even numbers. The standard measure
of a computation at the time was the number of multiplications it required.
Researchers in AI were proud of the fact that there were no multiplications
at all in their programs, though these programs were complex enough to
prove theorems or play games. The issue was actively pursued as a struggle
over how the computer was to be viewed. However, it was joined in an
asymmetric way. The bulk of the computer field, and all its responsible
opinion-leaders, simply adopted the view that computers are number
manipulators. There was no attempt to argue against the view that com¬
puters are symbol manipulators. It was just ignored, and the standard inter¬
pretation maintained. Researchers in AI, on the other hand, were actively
eng&ged in promoting the new view, considering the standard one to be a
radical misreading of the nature of the computer and one that provided a
significant barrier to the view that computers could be intelligent.
The result ot this clash of views was to isolate AI within computer sci-
INTELLECTUAL ISSUES IN THE HISTORY OF Al 197
ence. AI remained a part of computer science, but one with a special point of
view that made it somewhat suspect, indeed somewhat radical. This isola¬
tion is important historically, for it has affected the professional and disci¬
plinary organization of the two fields. It derives ultimately, no doubt, from a
basic divergence of views about whether computers can or cannot exhibit
intelligence. This overarching issue, of course, continued to be important on
its own, as witnessed by the debates that occurred throughout the 1950s on
whether machines could think. But the more specific issues that it spawned
also had independent lives.
The issue of symbols versus numbers did not arise until after the first AI
programs came into existence, circa 1955. Before that time, programs were
classified as numerical versus nonnumerical. This latter class was a miscel¬
lany of all the things that processed data types other than numbers—
expressions, images, text, and so forth.2 This included the few game-playing
and logic programs but much else as well. The symbols-versus-numbers
issue emerged only when a positive alternative became formulated, that is,
symbolic manipulation. This was not a synonym for nonnumerical process¬
ing, for it laid the groundwork for the separation of image- and text¬
processing from AI. Indeed, the work on machine translation, which started
in the early 1950s, was initially considered as one strand in the development
of intelligence on machines. [Locke and Booth, 1957.] But that effort be¬
came concerned with text and not symbols and developed its own identity as
computational linguistics. (All of this, of course, was before text processing
in its current meaning emerged—an event that bore no significant relation to
the development of computational-linguistics.)
I have placed the ending of this issue at about 1965, although I do not have
a significant marker event for its demise. The issue is certainly not alive now
and has not been for a long time. In part, this is due to the prominence of
many nonnumerical data types in computer science generally, such as text
and graphics. These make the characterization of computers as number
manipulators no longer ring true. In part, it is due to the shift within theoret¬
ical computer science to algebraic and logical formalisms, with the concur¬
rent retreat of numerical analysis from its early dominant role. In part, of
course, it is due to the success of AI itself and the demonstrations it brought
forward of the symbolic character of computation. It is tempting to say that
the cause was simply the growth of scientific understanding—but such rea¬
sons do not fare well in historical accounts. In any event, my recollection is
that the symbols/numbers issue was no longer prominent by the late 1960s,
though a little historical digging might place it five years later.
2The concept of data type did not arrive in clear form until much later.
198 ALLEN NEWELL
scientists who believed that logic was the appropriate class of systems? And
if not, why not? First, by logical systems is meant the class of systems that
do logical operations, such as AND, OR, NOT, and so forth.4 This is the
class corresponding to the logic level in the hierarchy of computer struc¬
tures. The logic level is located between the circuit level and the program
(symbol) level. All three levels are equally comprehensive and provide three
possibilities for ways of describing intelligent systems. Indeed, circuit and
program levels correspond exactly to the continuous and symbol positions
of the issue under discussion. Now, in fact, in the early days, there were
attempts to build logic machines and discuss the behavior of systems di¬
rectly in terms of logic circuits. The classical neural networks of McCulloch
and Pitts were an effort at modeling the neural system at the logic level.
[McCulloch and Pitts, 1943.] But all these efforts rapidly died out and were
all but gone by the mid-1960s. My own guess about why this happened is that
the hierarchy of computer levels indicated quite clearly what to do with a
logic level—namely, compose a higher level system. But this implied simply
reproducing existing program-level systems, at least without some new or¬
ganizational ideas at the program level. But the logic level provided no such
ideas, nor could it. Thus, there was nowhere to go. In fact, the history of
these efforts seems quite obscure, and tracing the demise of logic as a
system language for intelligent systems would be a substantial, though re¬
warding, undertaking.
4 It might also mean the class of theorem-proving systems using logical calculi; but this is really
a subclass of symbol systems.
200 ALLEN NEWELL
suffering the same fate as logic systems. Whereas with logic systems there
was nowhere to go except toward program-like organizations, with continu¬
ous systems there was the brain to model. We need not demand an answer to
what the higher organization would be, we could just take as guide the brain
as revealed in current neurophysiological work. It is true, of course, that in
the late 1940s and early 1950s, the discrete approximation to the nervous
system (neurons as digital threshold devices) promised to provide neuro¬
physiological inspiration for the class of logic systems. But under a barrage
of criticism, even the engineers came to accept the nervous system as too
complex to be modeled by logic-level systems, which is to say, its con¬
tinuities had to be taken seriously. Thus, without any source of inspiration,
logic-level systems faded away as a separate language for modeling intelli¬
gence, but continuous systems remained.
Yet another issue can be identified that is coordinated with the issue of
symbolic versus continuous systems. AI concentrated on creating perform¬
ance systems, that is, systems that performed some task demanding intelli¬
gence. Cybernetics and pattern-recognition research concentrated on creat¬
ing systems that learned. Indeed, another subfield grew up that called itself
self-organizing systems. [Yovits, Jacobi, and Goldstein, 1962.] In practice,
self-organizing systems largely overlapped with the work in pattern recogni¬
tion and it had common roots in cybernetics. But self-organizing systems
took the problem of learning as the central focus rather than the problem of
recognition. For instance, within self-organizing systems, there was consid¬
erable interest in embryology, even though it had little to do with recognition
at the time.
Through the early 1960s, all the researchers concerned with mechanistic
approaches to mental functions knew about each other’s work and attended
the same conferences. It was one big, somewhat chaotic, scientific happen¬
ing. The four issues I have identified—continuous versus symbolic systems,
problem-solving versus recognition, psychology versus neurophysiology,
and performance versus learning—provided a large space within which the
total field sorted itself out. Workers of a wide combination of persuasions on
these issues could be identified. Until the mid-1950s, the central focus had
been dominated by cybernetics, which had a position on two of the issues—
using continuous systems and orientation toward neurophysiology—but no
strong position on the other two. For instance, cybernetics did not concern
itself with problem-solving at all. The emergence of programs as a medium
of exploration activated all four of these issues, which then gradually led to
the emergence of a single composite issue defined by a coordination of all
four dimensions. This process was essentially complete by 1965, although I
do not have any marker event. Certainly by 1971, at the second International
Joint Conference on Artificial Intelligence in London, it was decided that
202 ALLEN NEWELL
It is worth noting for future reference that most pattern-recognition and self¬
organizing systems were highly parallel network structures. Many, but not
all, were modeled after neurophysiological structures. Most symbolic-
performance systems were serial programs. Thus, the contrast between se¬
rial and parallel (especially highly parallel) systems was explicit during the
first decade of AI. The contrast was coordinated with the other four issues I
have just discussed. However, I do not recollect it playing nearly as active a
role as any of the other four, so I have simply added it on as a comment.
These issues were not the only ones that emerged in the first decade of AI’s
existence, nor the most important. A candidate for the most important initial
issue was AI’s development of heuristic programs in contradistinction to
algorithms. Algorithms were taken to be programs that guaranteed that they
would solve a problem or solve it within given time bounds. Good programs
were algorithmic, and if not, the fault lay with the programmer, who had
failed to analyze his or her problem sufficiently—to know what the program
should do to solve this problem. Heuristic programs, on the other hand,
were programs that operated by means of heuristic rules of thumb-
approximate, partial knowledge that might aid in the discovery of the solu¬
tion but could not guarantee to do so. The distinction implied that intelligent
problem-solving could be attained by heuristic programs. For a short while,
one name for the field of AI was heuristic programming, reflecting, in part, a
coordination with such subfields as linear programming and dynamic pro¬
gramming (which were also just then emerging).
An important effect of this issue was to isolate AI within computer sci¬
ence but along a different dimension than the issue of symbols versus num¬
bers. Heuristic programming indicates a commitment to a different course
than finding the best engineering solution or mathematical analysis of a
problem. According to the standard engineering ethos, the proper use of the
computer requires the engineer or analyst to exert his or her best intellectual
efforts studying the problem, find the best solution possible, and then pro¬
gram that solution. Providing a program with some half-baked, unanalyzed
rules seemed odd at best and irrational, or even frivolous, at worst. A good
example ot this tension can be found in the work of Wang, whose theorem¬
proving program performed much better than the LOGIC THEORIST.
[Newell, Shaw, and Simon, 1957; Wang, I960.] The thrust of Wang’s posi¬
tion was that much better theorem-provers could be built if appropriate
results in mathematical logic were exploited. The defense by the AI commu-
INTELLECTUAL ISSUES IN THE HISTORY OF Al 203
nity stressed finding how humans would solve such problems, in effect deny¬
ing that the fullest analysis of experimental tasks was the object of the
investigation. Another important example was the MACSYMA project to
construct an effective computer system for physicists and engineers to do
symbolic manipulation of mathematical expressions. Although this work
grew out of two prior efforts in AI, it was cast by its leaders as “not part of
AI,” but, rather, as part of an area of computer science called symbolic
manipulation, which took a thoroughgoing engineering and analytical at¬
titude. [Slagle, 1963; Moses, 1967.]
I have put the demise of the issue at the mid-1960s; the issue gradually
ceased to be discussed, though the distinction continues to be made in
textbooks and introductory treatments. Once the field was underway, with
lots of AI systems to provide examples, the point at issue became transpar¬
ent. Moreover, the distinction has difficulty in being transformed into a
technical one, because it is tied to features external to the procedure itself,
namely, to the problem that is supposed to be solved and the state of knowl¬
edge of the user of the procedure.
5 In fact, the first report of FORTRAN at a scientific meeting occurred at the same session as the
first report of a list-processing language. [Backus et al., 1957; Newell and Shaw, 1957 ]
204 ALLEN NEWELL
and-run operating philosophy. Just for fun, I have indicated the end of this
issue in 1985, on the assumption that the coming generation of powerful
personal computers will finally force all languages to come to terms with full
dynamic capabilities in order to permit interactive programming. But this is
pure conjecture, and the separation may now be wide enough to require a
generation to heal.
The grounds for this issue can be traced to demands for efficiency on the
one hand versus demands for flexibility on the other; perhaps the issue
should have been so labeled. For instance, the main programming commu¬
nity in the late 1950s also had a strong negative reaction to list-processing,
because of its giving up half the memory just to link the actual data together.
But, although the general efficiency issue was always on the surface of
discussions, the total situation seems better described in terms of distinct
structural alternatives, that is, interpreters versus compilers, list structures
versus arrays, and recursion versus iteration.
One issue that surfaced right from the start of AI was whether to make
machines be intelligent by simulating human intelligence or by relying on
engineering analysis of the task. Those who were primarily trying to under¬
stand human intelligence inclined naturally to the simulation view; those
who were primarily engineers inclined to the pure task-analysis view. The
principle was frequently invoked that we do not build a flying machine by
simulating bird flight. On the simulation side, there was more than one
position. The majority took the view that casual observation and casual
introspection was the appropriate approach—that is, the human was a
source of good ideas, not of detail. A few, usually with strong psychological
interests or affiliations, took the view that actual experimental data on hu¬
mans should be examined.
This issue seems never to have produced any important crises or changes
of direction in the field; however, it has probably decreased the amount of
mutual understanding. There seems to be little movement in a scientist’s
position on this issue. Each investigator finds his or her niche and stays
there, understanding only superficially how those with different approaches
operate. The position adopted probably reflects fairly deep attitudes, such as
determine whether a scientist goes into an engineering discipline or a social/
behavioral discipline in the first place. This is to be contrasted with many
fields where methods are effectively neutral means to ends, to be used by all
scientists as the science demands. There is little indication of diminution of
this issue over the years, although starting in the 1970s, there has been some
increase in the general use of protocols to aid the design of AI systems, even
when there is no psychological interest.
This completes the set of new issues that arose coincident with the birth
of AI. Five of them—symbolic versus continuous systems, problem-solving
INTELLECTUAL ISSUES IN THE HISTORY OF Al 205
An issue that surfaced about five years after the beginning of AI was whether
the proper objective was to construct systems that replace humans entirely
or to augment the human use of computers. The fundamentally ethical di¬
mension of this issue is evident. Yet, it was not overtly presented as an issue
of social ethics but, rather, as a matter of individual preference. An inves¬
tigator would simply go on record one way or another, in the prefaces of his
or her papers, so to speak. Yet, there was often an overtone, if not of ethical
superiority, of concordance with the highest ideals in the field. Those whose
inclinations were toward AI did not so much meet this issue head on
as ignore it. Indeed, it was perfectly possible to take the view that work
in AI constituted the necessary exploration for man/computer symbiosis.
[Licklider, I960.]
A relatively weak issue such as this could not really become established
unless man/machine cooperation offered technical possibilities and chal¬
lenges as exciting as constructing intelligent machines. Thus, the beginning
of this issue coincides with the appearance of interesting interactive sys¬
tems, such as SKETCHPAD, which had an immense influence on the field.
[Sutherland, 1963.]
Artificial intelligence scientists have had a relatively large involvement in
the development of user/computer interaction throughout the history of
computer science; for example, in time-sharing in the 1960s and 1970s, in
making languages interactive in the 1970s, and in developing personal ma¬
chines in the early 1980s. One explicit justification given for this involvement
was that AI itself needed much better programming tools to create intelligent
programs—a reason quite independent of the issue presented heie. How¬
ever, it is not possible to untangle the relations between them without some
rather careful historical analysis.
Many of those who opted for working in user/computer cooperation
tended not to become part of AI as the latter gradually evolved into a field.
However, as I have already noted, it was entirely possible to work in both AI
and user/computer cooperation. Still, the net result was an additional factor
of separation between those in AI and those in neighboring parts of com¬
puter science.
206 ALLEN NEWELL
It is easy to distinguish the knowledge that an intelligent agent has from the
procedures that might be necessary to put that knowledge to work to exhibit
the intelligence in action.6 The initial period in AI was devoted almost exclu¬
sively to bringing into existence modes of heuristic processing worthy of
consideration. In 1959, John McCarthy initiated a research position that
distinguished such study sharply from the study of appropriate logical for¬
malisms to represent the full range of knowledge necessary for intelligent
behavior. [McCarthy, 1959.] This study was clearly that of epistemology—
the study of the nature of knowledge. It bore kinship with the subfield of
philosophy by the same name, although, as with so many other potential
connections of AI and philosophy, the orientation of the two fields is highly
divergent, although the domain of interest is nominally the same.
There has been little controversy over this issue, although the two poles
led to radically different distributions of research effort. Work on epistemol¬
ogy within AI has remained extremely limited throughout, although recently
there has been a substantial increase. [D. G. Bobrow, 1980.]
In the first years of AI, through the early 1960s, AI programs were character¬
ized simply as highly complex programs, without any particular notion of
common structure. For instance, the field was also called complex informa¬
tion processing as well as heuristic programming. By 1965, however, it had
become clear that the main AI programs used the same fundamental tech¬
nique, which became known as heuristic search. [Newell and Ernst, 1965.]
This involves the formulation of the problem to be solved as combinatorial
search, with the heuristics cast in specific roles to guide the search, such as
the selection of which step to take next, evaluation of a new state in the
space, comparison of the present state to the posited goal-state, and so on.
As the scope of AI programs seemed to narrow, there arose a belief in some
AI scientists that the essence of intelligence lay not in search, but in large
amounts of highly specific knowledge, or expertise. This issue was well
enough established by the mid-1970s to occasion the declaration that a
paradigm shift in AI had already occurred, the original paradigm having been
heuristic search with little knowledge of the task domain and the new
paradigm being knowledge-intensive programs. [Goldstein and Papert,
1977.]
It may be doubted that these changes amounted to an actual paradigm
shift. What clearly did happen was a major expansion of AI research to
Said this way, the connection of this issue to the competence/performance issue discussed
later would seem to be overwhelming. However, the research programmes associated with the
two issues have never made common cause.
INTELLECTUAL ISSUES IN THE HISTORY OF Al 207
Another way to characterize the major early AI programs is that they took a
single well-defined difficult task requiring intelligence and demonstrated that
a machine could perform it. Theorem-proving, chess and checkers playing,
symbolic integration, IQ-analogy tasks, and such management-science tasks
as assembly-line balancing—all these fit this description. Again, there was a
reaction to this. Although AI could do these sorts of tasks, it could not do the
wide range of presumably trivial tasks we refer to as having common sense.
The need was for generality in AI programs, not power.
This call had been issued early enough. [McCarthy, 1959.] However, it
was really not until the mid-1960s that a significant shift occurred in the field
toward the generality and commonsense side. This gave rise to using small
constructed puzzles and artificial problems to illustrate various components
of everyday reasoning. A typical example was the monkey-and-bananas
task, patterned after simple tasks solved by Kohler’s chimpanzee, Sultan.
Whereas such problems would have seemed insignificant in the early years,
they now became useful, because the goal of research was no longer power,
but understanding how commonsense reasoning could occur.
By 1975, this shift had run its course, and new concerns for working with
relatively large-scale real problems took over with the development ot ex¬
pert systems already mentioned. As could have been expected, the end of
this period of emphasis did not mean a shift back to the original issue.
Although expert systems tackled real problems and, hence, were obviously
powerful, they did not achieve their power by the heuristic-search tech¬
niques of the early years; instead they used large amounts of domain-specific
knowledge (coupled, sometimes, with modest search).
208 ALLEN NEWELL
The Chomskian revolution in linguistics also started in the late 1950s. It was,
along with AI, just one of many similar and interrelated developments in
engineering, systems, and operational analysis. Although each of these de¬
velopments had a particularly intense significance for some particular field,
for example, linguistics or computer science, they all formed a common
interdisciplinary flux. Gradually, these activities sorted themselves into sep¬
arate subfields or disciplines, developing opposing positions on the issues
previously laid out, as we have seen for AI vis-a-vis cybernetics and pattern
recognition.
In many ways, linguistics was a special case. It was already a well-formed
discipline, and the revolution was at the heart of the discipline, not in some
peripheral aspect that could have split off and aligned with other intellectual
endeavors. Furthermore, only very few linguists participated in the general
flux that was occurring in the world of engineering and applied mathematics.
Linguistics was culturally and organizationally quite distinct, having strong
roots in the humanities. In fact, it probably made an immense difference that
Noam Chomsky became affiliated with the Massachusetts Institute of Tech¬
nology (MIT).
It was not until the mid-1960s that issues emerged that determined rela¬
tions between linguistics and other subfields and disciplines. A principal
issue was the distinction between competence and performance, which was
moved to a central position in the new linguistics by Chomsky. [Chomsky,
1965.] Linguistic competence was the general knowledge a speaker had of
the language, in particular, of the generative grammar of the language. Per¬
formance was the actual production of utterances, which could be affected
by many additional factors, such as cognitive limits, states of stress, or even
deliberate modifications for effect. The distinction made useful operational
sense for linguistics, because there were two sources of evidence about
human-language capabilities, the actual utterance and the judgment of gram-
maticality—a sort of recall/recognition difference, although that analogy
was never exploited.
This distinction might seem innocuous from the standpoint of science
history, that is, purely technical. In fact, it served to separate quite radically
the sciences concerned primarily with performance, namely AI, computa¬
tional linguistics, cognitive psychology, and psycholinguistics, from linguis-
INTELLECTUAL ISSUES IN THE HISTORY OF Al 209
tics proper. Linguistics itself declared that it was not interested in perform¬
ance. More cautiously said, competence issues were to have absolute
priority on the research agenda. But the effect was the same: Work in any of
the performance fields was basically irrelevant to the development of lin¬
guistics. There could be a flow from linguistics to these other fields, and,
indeed, there was an immense flow to psycholinguistics, but there could not
be any significant flow in the other direction.7
A more effective field-splitter would be hard to find. It has remained in
effect ever since, with the competence/performance distinction being ex¬
tended to other domains of mentality. This has certainly not been the only
significant cause of the separateness of AI from linguistics. There are impoi-
tant isolating differences in method, style of research, and attitudes toward
evidence. Many of these other issues share substance with the competence/
performance distinction and affect the separation between psychology and
linguistics much more than that between AI and linguistics. Thus, perhaps
these issues can be left to one side.
7This is not the whole story of the relations of linguistics with other fields; for example, there
have been important contacts with logic and philosophy.
210 ALLEN NEWELL
In 1965, AI took back the problem of recognition that had become the
intellectual property of the pattern-recognition community. This can be
marked rather precisely by the work of Roberts on the recognition of three-
dimensional polyhedra. [Roberts, 1965.] The essential features were two:
First, recognition was articulated, that is, the scene had to be decomposed
or segmented into subparts, each of which might need to be recognized to be
a different thing. Thus, the result of recognition was a description of a scene
rather than just an identification of an object. But a description is a symbolic
Although the term architecture is just now coming into common use in psychology.
211
INTELLECTUAL ISSUES IN THE HISTORY OF Al
structure that has to be constructed, and such processes were quite outside
the scope of the pattern-recognition techniques of the time, though exactly
of the sort provided by AI. Second, a major source of knowledge for making
such recognitions came from adopting a model of the situation (e.g., it con¬
sists only of polyhedra). This made recognition processes strongly inferen¬
tial, again fitting in well with work in AI, but not with work in pattern
recognition.
By the late 1960s, work on vision was going on throughout AI, but the
transformation went further than just vision. Three laboratories (at MIT,
Stanford, and the Stanford Research Institute) started major efforts in robot¬
ics. Vision was to be coupled with arms and motion and in at least one AI
center (Stanford), with speech. The entire enterprise was radically different
in its focus and problems from the research in pattern recognition that was
still going on in parallel in departments and research centers of electrical
engineering. In fact, there was little actual controversy to speak of. Both
groups simply did their thing. But likewise, there was no substantial rap¬
prochement.
9There was work on phonology, but the domain lay outside the range of interest of AI and. in
fact, of psychology as well.
212 ALLEN NEWELL
Recall that resolution theorem-proving flourished in the late 1960s and bid
fair to become the engine at the center of all reasoning. In fact, it took only a
few years for the approach to come up against its limitations. Despite in¬
creases in power, relative to prior efforts, theorem provers were unable to
handle any but trivial tasks. Getting from logic to real mathematics—seen
always as a major necessary hurdle—seemed as far away as ever.
The reaction to this state of affairs became known as the procedural/
declarative controversy. Theorem provers were organized as a large homo¬
geneous database of declarative statements (clauses in resolution), over
which an inference engine worked to produce new true statements to add to
the database. This was the essence of a declarative representation of knowl¬
edge and its attractions were many. Its difficulty lay in the costs of process¬
ing. The inference engine treated all expressions in the database alike or,
more precisely, without regard for their semantics. There also seemed no
10Among the contributors to the first conference on Theoretical Issues in Natural Language
Processing, a series that became the forum tor this subfield, I can identify only one mainstream
linguist. [Schank and Nash-Webber, 1975.]
INTELLECTUAL ISSUES IN THE HISTORY OF Al 215
way for a theorem prover to be given information about how to solve prob¬
lems. These two features added up to a major combinatorial explosion. The
remedy—the procedural side of the issue—lay (so it was claimed) in encod¬
ing information about the task in procedures. Then knowledge would be
associated directly with the procedures that were to apply it; indeed, the
procedures would embody the knowledge and, thus, not have to be inter¬
preted by another inference engine. This would permit the appropriate guid¬
ance for problem-solving and, thus, keep the combinatorial explosion under
control.
There are irremediable flaws in both sides of the argument whethei
knowledge should be coded in procedural or declarative form, just as there
are irremediable flaws in both sides of the argument whether a program is
heuristic or algorithmic. Both procedural and declarative representations are
necessary to make any computation at all happen. In consequence, argu¬
ments over the issue were largely inconclusive, although they produced the
closest thing to a public issue-controversy in AI’s short history. However,
the effect on the course of AI research was enormous. First, work on
theorem-proving shrank to a trickle, with what remained mostly devoted to
nonresolution theorem-proving. Second, so-called planning languages
emerged as a result—PLANNER, QA4, CONNIVER, POPLAR, and so
forth. [Bobrow and Raphael, 1974.] These programming-language systems
were intended to provide a vehicle for writing the sorts of domain-
dependent, procedure-oriented theorem provers called for in the debate.
While that did not quite happen, these languages in themselves provided a
major conceptual advance in the field. The effects of this issue had about run
their course by 1980.
11 This is because list structures approximate general symbolic systems. The neutrality is easily
confirmed in the continued and universal use of list-processing languages to realize systems o
all kinds along this dimension.
216 ALLEN NEWELL
proving in the late 1960s brought with it as a side effect the clause as the unit
of representation. The clause was a primitive assertion that could not be
broken down into a conjunction of other assertions—primitive predicates P,
negations of primitive predicates ~P, disjunctions P or Q, implications P
implies Q, and so forth. The total knowledge of the system was to be repre¬
sented as the conjunction of clauses—that is, to use the old Gestaltist
phrase, as an And-sum of separate bits of knowledge.
Thus, the issue of size of representational unit grew out of the same
ground as the procedural versus declarative controversy, and, indeed, it was
articulated by the same group at MIT who had made most of the latter issue.
As is always the case, concern was, in fact, widespread but had been subor¬
dinated to other concerns. [Abelson, 1973; Norman, 1973; Schank, 1973.]
Minsky was the first one to give clear voice to the concern. The effect of the
paper was dramatic, despite the fact that the paper itself was entirely specu¬
lative and discursive. Throughout AI, the concept of the frame as the appro¬
priate data structure was widely embraced. By 1980, frame systems were an
established part of AI, and a very substantial fraction of the work in knowl¬
edge representation was involved in such systems.
Much follows on this development (in conjunction with the procedural/
declarative issue)—the rise of substantial research effort in knowledge
representation and the strengthening of renewed ties with philosophy.
[Brachman and Smith, 1980.] These efforts conjoin with those of AI epis¬
temology, discussed earlier. They raise some new issues, such as the rela¬
tion of philosophic work on meaning to directly inspired computational mod¬
els. But these issues have not yet jelled enough to be included in their own
right.
Another example is the problem of induction, where philosophy is concerned with the cer¬
tainty ot induction and AI is concerned with performing the inductions. [Newell, 1973c.]
INTELLECTUAL ISSUES IN THE HISTORY OF Al 217
The issue has not been especially active, but it has been raised.
[Gunderson, 1971.] It is argued that performance functions (i.e., those func¬
tions AI currently deals with, called program-receptive functions) can be
mechanized; but that sentient functions (i.e., feelings, called program-
resistant functions) cannot. Whether this will ever grow to a substantial
controversy is hard to tell at this point. It is certainly available as a reserve
position that can serve to separate AI from the philosophy of mind. It adds to
the general background concern, discussed in the first occurrence of this
issue, of the absence of emotion and feeling in the development of intelligent
systems.
As noted in the power/generality issue, the field took a shift in the mid-1960s
away from powerful programs toward programs that could exhibit common
sense. Further, as noted in the language/tasks issue, this line further trans¬
muted to being concerned with understanding via the understanding of natu¬
ral language. Concomitantly, programs were often built to work on small
simple illustrative tasks or environments, usually puzzles or made-up
situations.
By the mid-1970s some systems had been developed that worked with real
tasks that had substantial intellectual content, to judge from their role in the
real world. The initial such system can be taken to be DENDRAL, which
determined the structural formula for chemical molecules, given the data on
the mass spectrogram.13 [Lindsay, Buchanan, Feigenbaum, and Lederberg,
1980.] DENDRAL began in the late 1960s and grew in power throughout the
early 1970s. It was joined in the mid-1970s by several systems that per¬
formed competently in real medical-diagnosis tasks, of which MYCIN was
the paradigm. [Shortliffe, 1974.] This was the immediate locus of expert
systems, which, as previously noted, grew up as part of the general empha¬
sis on knowledge in contrast to search. With it grew an attitude that AI in
general should no longer work on small illustrative, artificial tasks but that it
was time to work on real tasks. The simple artificial tasks came to be called
toy tasks, not just because the term conveys the contrast between childish
and grown-up pursuits, but also because stacking children’s blocks had be¬
come a favorite illustrative task environment.
The tension between basic research and application exists in all sciences
at all times. Sciences sometimes build institutional structures to contain the
tension. As we saw in the issue of science versus engineering, computer
science has kept its basic and applied components mixed together in a single
discipline, thus exacerbating the tension. The tension was, in fact, especially
13The other system often mentioned similarly is MACSYMA, the highly sophisticated program
at the Massachusetts Institute of Technology for doing symbolic mathematics. As mentioned
earlier, it had deliberately removed itself from being an AI program.
218 ALLEN NEWELL
severe for AI during the decade of the 1970s. The climate in Washington was
not benign for basic research in general, and there was sustained pressure
from AI’s primary government funding agency (DARPA—Defense Ad¬
vanced Research Projects Agency) to make AI pay off. That said, however,
the distinction between toy versus real tasks is not solely the distinction
between basic and applied research. Tasks taken from the real world and
performed by intelligent humans as part of their working lives carry a prima
facie guarantee of demanding appropriate intelligent activity by systems that
would perform them. It can be argued that such tasks are the appropriate
ones for AI to work on, even if the goal is basic research. Thus, the toy-
versus-real-tasks issue stands ambiguously for both meanings—basic versus
applied and irrelevant versus relevant basic science.
By the mid-1970s, computer science had for some time been seriously ex¬
ploring multiprogramming and multiprocessing. These provided the ground¬
work for considering parallel systems for doing AI. A major instigation
occurred with the development of the Hearsay-II model of speech under¬
standing. [Lesser and Erman, 1977.] Hearsay-II comprised a number of
knowledge sources (acoustic, phonetic, phonological, lexical, syntactic, se¬
mantic, and pragmatic), each working concurrently and independently off a
common blackboard that contained the current working state about the ut¬
terance and each contributing their bit to the evolving recognition and react¬
ing to the bits provided by the others.
The Hearsay-II structure was certainly a parallel one, but it was at a level
of parallelism quite different from earlier network models, namely, a modest
number (tens) of functionally specialized processes. Furthermore, individual
processes remained fundamentally symbolic (even though lots of signal¬
processing was inherent in the speech-recognition task). Hearsay-II was
only one of several efforts to pursue the notion that an intelligent system
should be thought of in terms of communicating subprocesses rather than as
an individual serial machine. A metaphor arose for thinking about an intelli¬
gent system—the scientific community metaphor—which took the opera¬
tion of science, with its notion of cooperation, publication, experiment,
criticism, education, and so forth, as the appropriate model for intelligent
activity. Gradually, a group of people emerged interested in working on
distributed AI.
they hold the key to intelligence, that each learning program that was con¬
structed received substantial attention.14 [Samuel, 1959; D. A. Waterman,
1970; Winston, 1970; Sussman, 1975.] However, each learning system was
relatively idiosyncratic, with its own interesting lessons, so that the whole
did not add up to a coherent effort for the field.
A reversal of this state of affairs developed by the late 1970s. It was
triggered by the spread of a class of programming systems, called produc¬
tion, or rule-based systems, which are used for both constructing expert
systems and analyzing human cognition. [Waterman and Hayes-Roth, 1978.]
To appreciate their role in the resurgence of work on learning, we must take
a step back. To create a learning system requires solving two research
problems. First, a space of potential performance programs must be created,
in which learning will constitute moving from one program to another,
searching for programs with better performance. If the space of programs is
too vast and irregular, then learning is, in effect, automatic programming,
and it becomes extremely difficult. If the space is too limited, then learning is
easy, but the performance programs are of little significance. Determining
the right space is, thus, a critical research activity. Second, given the space,
it is still necessary to design an interesting learning system, for the space
only lays out the possibilities. Thus, inventing the learning system is also a
critical research activity. A major reason why early AI learning-systems
seemed so idiosyncratic was that each made unique choices on both these
dimensions. Most important, doing research on learning was doing a double
task and taking a double risk.
A production system is composed entirely of a set of if-then rules (if such
and such conditions hold, then execute such and such actions). At each
instant, the rules that hold are recognized, and a single rule is selected to
execute. In such a system, the natural space of performance programs con¬
sists of subsets of if-then rules, and the primitive act of learning is to add a
new rule to the existing set (or sometimes to modify an existing rule in some
simple way, such as by adding another condition). This space of perform¬
ance programs is neither too limited nor too open, since it is easy to restrict
the rules to be learned to a special class. As a consequence, the first research
choice is essentially made for the researcher, who can then concentrate on
constructing an interesting learning program. Moreover, learning programs
will have much in common, since they now use similar spaces of per¬
formance programs. Indeed, this is just what happened in the late 1970s
as researchers began to construct a wide variety of small learning systems,
all built around variants of the production-system formalism. [Michalski,
Carbonell, and Mitchell, 1983.] It must be realized, of course, that such fo¬
cusing of effort does not remove the collective risk. If production systems
14Some other systems were built, which might have been viewed as learning systems, but,
instead, were taken simply to be performance programs in specialized task environments, for
example, induction programs.
220 ALLEN NEWELL
are the wrong program organization to be exploring, then the entire field is
moving down an unproductive path.
Robotics has returned to AI after having left it for most of the 1970s. Perhaps
it is unfortunate to call the issue problem-solving versus recognition, since
recognition is only one aspect of robotics. The main sources of the new wave
of effort are external to AI—industrial robotics plus the concern of the
decline in American productivity and the trade position of the United States
vis-a-vis Japan and West Germany. The initial growth of industrial robotics
took place largely outside of AI as a strictly engineering endeavor. As a
result, the initial growth tended to minimize the intelligence involved, for
example, sensory-motor coordination. One component of the new associa¬
tion of robotics with AI is the coupling of significant amounts of vision with
manipulators, reflecting the continued advance of vision capabilities in AI
throughout the 1970s. (Touch and kinesthetic sensing is increasingly impor¬
tant, too, but this does not build so strongly on prior progress in AI.) Impor¬
tantly, along with industrially motivated aspects, there is also a revival of
basic research in manipulation and movement in space and over real ter¬
rains.
It might seem that this is just another purely technical progression. But
with it has returned, as night follows day, the question of the relation of AI
and robotics as disciplines, just as the question was raised in the issue of
problem-solving versus recognition during the late 1960s. Is robotics a cen¬
tral part of AI or only an applied domain? Do graduate students in AI have to
understand the underlying science of mechanics and generalized coordinate
systems that are inherent in understanding manipulation and motion? Or is
that irrelevant to intelligence? Cases can be made either way. [Nilsson,
1982.]
PROLOG in the major AI centers in the United States, the issue takes on
additional coordinated dimensions. The outcome is far from clear at this
juncture.
DISCUSSION
It should be clear by now why I entered the caveats about historical accu¬
racy at the beginning. Each of the issues raises serious problems of charac¬
terization and historical grounding. No attempt has been made to define an
intellectual issue, so that some modestly objective way could be found to
generate a complete set of issues, for example, by placing a grid over the
literature of the field. Several additional issues might weil have emerged, and
some of those presented here might not have made the grade. Thus, the
population of issues exhibited must be taken, not just with a pinch of salt,
but soaked in a barrel of brine. Similar concerns attend dating the issues and
my interpretation of them; nevertheless, some comments about the total
picture seem worthwhile.
What Is Missing?
I do know why some issues did not make it. Three examples will illustrate
some reasons. The first is the broad but fundamental issue of the ethical use
of technology and the dehumanization of people by reduction to mechanism.
This issue engages all of technology and science. It seems particularly acute
for AI, perhaps, because the nature of mind seems so close to the quick. But
the history of science reminds us easily enough that at various stages as¬
tronomy, biology, and physics have seemed special targets for concern.
There has been continued and explicit discussion of these issues in connec¬
tion with AI. [Taube, 1961; Weizenbaum, 1976; McCorduck, 1979.] I have
not included them in the list of intellectual issues because they do not, in
general, seem to affect the course of the science. Where some aspect does
seem to do so, as in the issue of helping humans or replacing them, it has
been included. However, the broader issue certainly provides a thematic
background against which all work goes on in the field, increasing its ambi¬
guity, and the broader issue undoubtedly enters into individual decisions
about whether to work in the field and what topics to select.
The second example involves Hubert Dreyfus, who has been a persistent
and vocal critic of AI. [Dreyfus, 1972.] He has certainly become an issue for
the field; however, this does not necessarily produce an intellectual issue.
Dreyfus’s central intellectual objection, as I understand him, is that the
analysis of the context of human action into discrete elements is doomed to
failure. This objection is grounded in phenomenological philosophy. Unfor¬
tunately, this appears to be a nonissue as far as AI is concerned. The an¬
swers, refutations, and analyses that have been forthcoming to Dreyfus’s
INTELLECTUAL ISSUES IN THE HISTORY OF Al 223
AI versus Cybernetics
A cluster might seem to define a single underlying issue, which can then
replace component issues. However, the fact that issues are coordinated
does not make them identical. Some scientists can always be found who are
aligned in nonstandard patterns. In fact, some of the clusters seem much
more consistent than others. Thus, the multiplicity of issues keeps the
scientific scene complex, even though, because of clustering, it appears that
it should be clear and simple. In fact, many of the groupings are more easily
labeled by how they separate fields than by any coherent underlying concep¬
tual issue.
Clustering of issues does seem to be a common occurrence; for instance,
a standard advanced text on learning in psychology begins with a list of
seven dichotomous issues that characterize learning theories. [Hilgard and
Bower, 1948 and 1975, pp. 8-13.] The first three—peripheral versus central,
habits versus cognitive structures, and trial-and-error versus insight—form
a coordinated cluster that characterizes stimulus/response theories versus
cognitive theories (to which could even be added tough-minded versus ten¬
der-minded, the contrast William James used to distinguish the two main
types of psychologists). One possible source for such coordinated clusters is
the attempt to find multiple reasons to distinguish one approach from an¬
other. The approach comes first and the issues follow afterward. Then the
issues take on an autonomous intellectual life and what starts as rationaliza¬
tion ends up as analysis.
A major role of the issues here seems to be to carve up the total scientific
field into disciplines. AI, computer science, logic, cybernetics, pattern rec¬
ognition, linguistics, and cognitive psychology—all these seem to be dis¬
criminated in part by their position on these various issues. The issues, of
course, only serve as intermediaries for intellectual positions that derive
from many circumstances of history, methodological possibilities, and
specific scientific and technical ideas. Still, they seem to summarize a good
deal of what keeps the different fields apart, even though the fields have a
common scientific domain.
Is the large burst of issues that occurred at the birth of AI just an artifact
of my intent to gather issues for AI? If the period just before AI began, say
from 1940-1955, were examined carefully, would many more issues be
added? The relevant question should probably be taken with respect to some
other field as a base. Would a burst like this be found for cybernetics, which
started in 1940-1945? My own suspicion is yes, but I have not tried to verify
it.
Perhaps then the situation of AI could turn out to be typical. We would
find a plethora of issues in any science if we would but look and count; the
list from Hilgard and Bower might serve as a positive indicator. However,
before rushing to embrace this view, some counterevidence should be exam¬
ined. An interesting phenomenon in this same postwar period was the emer¬
gence of several one-theorem fields. Game theory, information theory,
226 ALLEN NEWELL
Another field, general systems theory, also had a single idea around which to build—that
there are common laws across all levels of systems from the atomic through cellular through
societal through astronomical. But there was no central result available, only the system view,
and this field has been markedly less successful than others in its growth and health.
INTELLECTUAL ISSUES IN THE HISTORY OF Al 227
CONCLUSION
Putting to one side questions about the accuracy of the particular set of
issues displayed in Table 1, of what use is a history of a scientific field in
terms of intellectual issues? To repeat once more: It cannot substitue for a
substantive history in terms of concepts, theories, and data; however, it
does seem to capture some of the flavor of the field in an era. It is clearly a
component of the paradigm of a field or of research programmes within a
field. And, let us confess it, intellectual issues have a certain spiciness about
them that makes them fun to talk and write about. Perhaps it is the sense of
touching fundamental issues. But perhaps it also echoes Bertrand Russell s
famous aphorism that dealing with intellectual issues has all the advantages
of theft over honest toil.
,
METHODOLOGICAL LINKS
BETWEEN ARTIFICIAL INTELLIGENCE
AND OTHER DISCIPLINES
Margaret A. Boden
some may feel that social psychology, for instance, has nothing to gain from
computational insights [e.g., Gauld and Shotter, 1977]; this should not be too
hastily assumed, however. Work within AI on the structure of action and the
attribution of intentions is relevant to theoretical discussions in social psy¬
chology. In general, AI supports the view that there may be generative rules
underlying social interaction or that social perception is a structured inter¬
pretative activity. But although these ideas are essentially consonant with a
computational viewpoint, specifying them in a particular case is a notori¬
ously difficult matter.
A general account of what sorts of psychological phenomena are or are
not grist for the AI mill would, of course, be very useful. But firm intellectual
ground could be provided for such an account only by a systematic theory of
representation. Philosophical discussions of the nature of intentionality are
clearly relevant [e.g., Fodor, 1981; Dennett, 1978/?]. Some philosophers
[e.g., Searle, 1980] argue that AI cannot model genuine (biological) inten¬
tionality, although discussions in recent issues of the peer-commentary jour¬
nal Behavioral and Brain Sciences show this claim to be highly controver¬
sial. But even John Searle admits that it can provide a scientifically useful
metaphor for intentionality. This is why AI is potentially relevant to studies
that are normally thought of as being humanistically oriented, such as social
and clinical psychology. [Boden, 1972.] Given that representational pro¬
cesses in computer models can function as heuristically fruitful analogues of
representational processes in our minds, the problem remains of providing
an account of the range and efficacy of such processes.
Artificial intelligence has shown that distinct representational forms affect
and effect inference in significantly different ways. Hinton’s work previ¬
ously mentioned is one of many examples that addresses such issues. An¬
other is Saul Amarel’s [1968] comparison of solutions to the “missionaries
and cannibals” problem, grounded in six representations of increasing
power; and a third is Aaron Sloman’s [1978b] discussion of analog represen¬
tations, which are interpreted by exploiting the similarity between their own
structure and that of the thing represented. However, there is—as yet—
little systematic understanding of the power and limitations of different rep¬
resentations. Work in computational logic is pertinent if it can show whether
or not a certain type of representation or computational process is, in princi¬
ple, capable of modeling a specific type of knowledge or simulating a given
class of psychological process.
General results in the philosophy of science apply to Al-based psychology
no less than to noncomputational theories. Some such results provide for a
rebuttal of common criticisms of the computational viewpoint. For instance,
even were it to turn out that AI is not appropriate for modeling many psycho¬
logical phenomena, we should not forget the Popperian point that we would
still have learnt something by the enterprise. Science involves conjecture
and refutation, and it is an advance to know that a specific conjecture has
been empirically rejected. Nor should we forget that some tricky
234 MARGARET A. BODEN
really representations and do not really possess intentionality but that these
terms as used by the computer scientist are parasitic on their use in the
human psychological context. The point is that the representational
metaphor (for such it is, in this view) is one that is suited to express psycho¬
logical phenomena (which alone are truly representational or intentional)
precisely because it is drawn from those parts of our everyday conceptual
scheme that concern these matters. For concepts to be fruitful in the theory
and methodology of an empirical psychology, it is not required that they be
interpreted as literal descriptions of the phenomena, just as we need not see
the atom as literally a solar system in order to benefit from the notion of
planetary electrons. So, whether computer programs specify representa¬
tional processes or merely “representational” ones, they are conceptually
close to hermeneutic forms of psychology rather than to those forms that
ignore subjectivity.
Educational projects within society at large are needed to alert people to
these facts, for most people associate computers with relatively stupid
brute-force programs (such as those used to calculate gas bills) and think of
them as machines and, therefore, as mechanistic. Most people fail to realize
that computational machines are radically different from noncomputational
machines and that they are not mechanistic in the sense that implies a denial
of subjectivity. The mistaken, though widespread, assumption that A1 mod¬
els of human beings are mechanistic in this sense may make people experi¬
ence a threat to—or even an undermining of—their personal autonomy and
moral responsibility. Behaviorism in psychology and the philosophy of mind
has been often, and justly, criticized for its underestimation or denial of
these psychological characteristics. But the computational approach, if
properly understood, is not open to such criticisms. To realize this is to
disarm the computational bogeyman.
ARTIFICIAL INTELLIGENCE
Cognition as Computation
Avron Barr
ARTIFICIAL INTELLIGENCE
technology that may have dramatic effects on our society. Already, experi¬
mental AI systems have generated interest and enthusiasm in industry and are
being developed commercially. These experimental systems include programs
that: solve some hard problems in chemistry, biology, geology, engineering,
and medicine at human-expert levels of performance; manipulate robotic
devices to perform some useful sensory-motor tasks; and answer questions
posed in restricted dialects of English (or French, Japanese, etc.). Useful AI
programs will play an important part in the evolution of the role of com¬
puters in our lives—a role that has changed in our lifetime from remote to
commonplace and if current expectations about computing cost and power
are correct, is likely to evolve further from useful to essential.
The intellectual currents of the times help direct scientists to the study of
certain phenomena. For the evolution of AI, the two most important forces
in the intellectual environment of the 1930s and 1940s were mathematical
logic, which had been under rapid development since the end of the
nineteenth century, and new ideas about computation. The logical systems
of Frege, Whitehead and Russell, Tarski, and others showed that some
aspects of reasoning could be formalized in a relatively simple framework.
much intelligence at all. But before these machines were even designed,
Church and Turing had seen that numbers were an inessential aspect of
computation—just one way of interpreting the internal states of the ma¬
chine.
The model of a Turing machine contains within it the notions both of what can
be computed and of universal machines—computers that can do anything that
can be done by any machine. (Newell and Simon, 1976u, p. 117.)
Turing, who has been called the father of AI, not only invented a simple,
universal, and nonnumerical model of computation, but also argued directly
for the possibility that computational mechanisms could behave in a way
that would be perceived as intelligent.
Thought was still wholly intangible and ineffable until modern formal logic
interpreted it as the manipulation of formal tokens. And it seemed still to
inhabit mainly the heaven of Platonic ideals, or the equally obscure spaces of
the human mind, until computers taught us how symbols could be processed by
machines. A.M. Turing . . . made his great contributions at the mid-century
crossroads of these developments that led from modern logic to the computer.
(Newell and Simon, 1976u, p. 125.)
As Allen Newell and Herbert Simon point out in the “Historical Epi¬
logue” to their classic work Human Problem Solving, there were other
strong intellectual currents from several directions that converged in the
middle of this century in the people who founded the science of artificial
intelligence. [Newell and Simon, 1972.] The concepts of cybernetics and
self-organizing systems of Wiener, McCulloch, and others dealt with the
macroscopic behavior of locally simple systems. The cyberneticists in¬
fluenced many fields, because their thinking spanned many fields, linking
ideas about the workings of the nervous system with information theory and
control theory as well as with logic and computation. Their ideas were part
of the Zeitgeist, but, in many cases, the cyberneticists influenced early
workers in AI more directly—as their teachers.
What eventually connected these diverse ideas was, of course, the devel¬
opment of computing machines themselves, conceived by Babbage and
guided in this century by Turing, von Neumann, and others. It was not long
after the machines became available that people began to try to write pro¬
grams to solve puzzles, play chess, and translate texts from one language to
another—the first AI programs. What was it about computers that triggered
240 AVRON BARR
This was part of the general insight of cybernetics, delayed by ten years and
applied to discrete symbolic behavior rather than to continuous feedback sys¬
tems. (Newell and Simon, 1972, pp. 869-870.)
new machines for performing tasks like solving equations and alphabetizing
lists—tasks that people perform using methods they can explicate. How¬
ever, people cannot specify in detail how they decide which move to make in
a game of chess or how they determine that two sentences mean the same
thing.
The realization that the detailed steps of almost all intelligent human
activity were unknown marked the beginning of artificial intelligence as a
separate part of computer science. AI researchers investigate different types
of computation, and different ways of describing computation, in an effort
not just to create intelligent artifacts, but also to understand what intelli¬
gence is. A basic tenet is that human intellectual capacity will best be de¬
scribed in the same terms as those invented to describe artificial intelligence
researchers’ programs. However, researchers are just beginning to learn
enough about those programs to know how to describe them scientifically—
in terms of concepts that illuminate the program’s nature and differentiate
among fundamental categories. These ideas about computation have been
developed in programs that perform many different tasks, sometimes at the
level of human performance, often at a much lower level. Most of these
methods are obviously not the same as those people use to perform the
tasks—some of them might be.
Problem Solving
The first big successes in AI were programs that could solve puzzles and
play games like chess. Techniques, such as looking ahead several moves and
242 AVRON BARR
dividing difficult problems into easier subproblems, evolved into the funda¬
mental AI techniques of search and problem reduction. Today’s programs
play championship-level checkers and backgammon, as well as very good
chess. Another problem-solving program that performs symbolic evaluation
of mathematical functions has attained very high levels of performance and
is being widely used by scientists and engineers. Some programs can even
improve their own performance with experience.
As discussed in the following sections, the open questions in this area
involve capabilities that human players exhibit but cannot articulate, such as
the chess master’s ability to see the board configuration in terms of meaning¬
ful patterns. Another basic open question involves the original conceptuali¬
zation of a problem, called in AI the choice of problem representation.
Humans often solve a problem by finding a way of thinking about it that
makes the solution easy; AI programs, so far, must be told how to think
about the problems they solve (i.e., the space in which to search for the
solution).
Logical Reasoning
Programming
Language
Learning
Certainly one of the most salient and important aspects of human intelli¬
gence is our ability to learn. However, this is an example of cognitive behav¬
ior that is so poorly understood that very little progress has been made in
achieving it in AI systems. Although there have been several interesting
attempts, including programs that learn from examples, from their own per¬
formance, or from advice from others, AI systems do not exhibit noticeable
learning ability.
Expert Systems
help the human expert locate the cause of errors in the system’s reasoning
when they occur.
Because I am most familiar with this area of AI research, and because its
imminent commercial applications are, I feel, indicative of important
changes in the field, much of the ensuing discussion of the role of AI in the
study of mind will refer to expert-systems research. The fact that these
systems represent vast amounts of knowledge obtained from human ex¬
perts; are used as tools to solve hard problems using this knowledge; can be
viewed as intermediaries between human problem-solvers; must explain
their thought processes in terms that people can understand; and are worth a
lot of money to people with real problems are the essential points that, I will
argue, will be true of all of AI someday—in fact, of computers in general—
and will change the role that AI research plays in the scientific study of
thought.
Open Problems
Although there has been much activity and progress in the twenty-five year
history of AI, some very central aspects of cognition have not yet been
achieved by computer programs. Our abilities to reason about others’ be¬
liefs; know the limits of our knowledge; visualize; be reminded of relevant
events; learn; reason by analogy; make plausible inferences; realize when
they are wrong, and know how to recover are not at all understood.
It is a fact that these and many other fundamental cognitive capabilities
may remain problematic for some time. But it is also a fact that computer
programs have successfully achieved a level of performance on a range of
intelligent behaviors unmatched by anything other than the human brain.
The failure of AI to achieve some seemingly simple cognitive capabilities in
computer programs becomes, according to the view of AI presented in this
paper, part of the set of phenomena to be explained by the new science.
machines without any prejudice toward making the system simple, biolog¬
ical, or humanoid.” (Minsky, 1963, p. 409.) But we do not start from scratch
in building the first program to accomplish some intelligent behavior; the
ideas about how that program is to work must come from somewhere. Fur¬
thermore, most AI researchers are interested in understanding the human
mind and actively seek hints about its nature in their experiments with their
programs.
The interest within AI in results and open problems of other disciplines
has been fully reciprocated by interest in and application of AI research
activity among researchers in other fields. Many experimental and theoret¬
ical insights in psychology and linguistics, at least, have been sparked by AI
techniques and results. Furthermore, I argue later on, this flow is likely to
increase dramatically in the future; its source is the variety of new phenom¬
ena displayed by AI systems—the number, quality, utility, and level of
activity of which will soon greatly increase. But first let us examine what
kind of interactions have taken place between AI and other disciplines in the
last 25 years.
. . . that body of experience which has grown up around the planning, evaluat¬
ing, and coding of complicated logical and mathematical automata will be the
focus of much of this information theory. . . .
it would be very satisfactory if one could talk about a “theory” of such auto¬
mata. Regrettably, what at this moment exists—and to what I must appeal—
can as yet be described only as an imperfectly articulated and hardly for¬
malized “body of experience.” (von Neumann, 1958, p. 2.)
Al: COGNITION AS COMPUTATION 247
We know shamefully little about our computers and their computations ....
We know very little, for instance, about how much computation a job should
require ....
When pioneering work by Newell, Shaw, and Simon, and other research
groups showed that “programming up” their intuitions about how humans
solve puzzles, find theorems, and so forth, was adequate for getting impres¬
sive results, the link between the study of human problem-solving and AI
research was firmly established.
Consider, for example, computer programs that play chess. Current pro¬
grams are quite proficient—the best experimental systems play at the human
expert level, but not so well as human chess masters. The programs work by
searching through a space of possible moves, that is, considering alternative
moves and their consequences several steps ahead in the game, just as
human players do. These programs, even some of the earliest versions,
could search through thousands of moves in the time it takes human players
to consider only a dozen or so alternatives. The theory of optimal search,
developed as a mathematical formalism (paralleling, as a matter of fact,
much work on optimal decision theory in operations research), constitutes
some of the core ideas of AI.
The reason that computers cannot beat the best human players is that
looking ahead is not all there is to chess. Since there are too many possible
moves to search exhaustively, even on the fastest imaginable computers,
alternative moves (board positions) must be evaluated without knowing for
sure which move will lead to a winning game, and this is one of those
abilities that human chess experts cannot make explicit. Psychological stud¬
ies have shown that chess masters have learned to see thousands of mean¬
ingful configurations of pieces when they look at chess positions, which
presumably helps them decide on the best move, but no one has yet sug¬
gested how to design a computer program that can identify these configu¬
rations.
Due to the lack of theory or intuition about human perception and learn¬
ing, AI progress on computer chess was virtually stopped, but it is quite
possible that new insights into a very general problem were gained. The
computer programs had pointed out more clearly than ever what kinds of
things it would be useful for a cognitive system to learn to see. It takes many
Al: COGNITION AS COMPUTATION 249
years for a chess expert to develop expertise, ability to understand the game
in terms of these concepts and patterns that he or she cannot explain easily,
if at all. The general problem is of course to determine what it is about our
experience that we apply to future problem-solving: What kind of knowledge
do we glean from our experience? The work on chess indicated some of the
demands that would be placed on this knowledge.
Ideas about getting computers to deal in some useful way with the human
languages, called natural languages by computer scientists, were conceived
before any machines were ever built. The first line of attack was to try to use
large bilingual dictionaries stored in the computers to translate sentences
from one language to another. [Barr and Feigenbaum, 1981, vol. 1, pp. 233-
238.] The machine would just look up the translation(s) of each word in the
original sentence, figure out its meaning (perhaps expressed in some interlin¬
gua), translate the words, and produce a syntactically correct version in the
target language.
It did not work. It became apparent early on that processing language in
any useful way involved understanding, which, in turn, involved a great deal
of knowledge about the world—in fact, it could be argued that the more we
know, the more we understand each sentence we read. And the level of
world knowledge needed for any useful language-processing is much higher
than our original intuitions led us to expect.
There has been a serious debate about whether AI work in computational
linguistics has enlightened us at all about the nature of language. [See
Dresher and Hornstein, 1976; and replies by Winograd, 1977; Schank and
Wilensky, 1977.] The position taken by AI researchers is that if our goal in
linguistics is to include understanding sentences like Do you have the time?
and We’ll have dinner after the kids wash their hands, which involve the
total relationship between the speakers, then there is much more to it than
the syntactic arrangement of words with well-defined meanings; that al¬
though the study in linguistics of systematic regularities within and between
natural languages is an important key to the nature of language and the
workings of the mind, it is only a small part of the problem of building a
useful language processor and, therefore, only a small part of an adequate
understanding of language.
For both people and machines, each in their own way, there is a serious
problem in common of making sense out of what they hear, see, or are told
about the world. The conceptual apparatus necessary to perform even a partial
feat of understanding is formidable and fascinating. (Schank and Abelson,
1977, p. 2.)
Linguists have almost totally ignored the question of how human understand¬
ing works. ... It has nevertheless been consistently regarded as important that
250 AVRON BARR
This idea that building AI systems requires the articulation of the detailed
nature of understanding, that is, that implementing a theory in a computer
program requires us to work out our fuzzy ideas and concepts, has been
suggested as a major contribution of AI research.
We are not oriented toward finding out which pieces of our theory are
quantifiable and testable in isolation. We feel that such questions can wait.
First we need to know if we have a viable theory .... (Schank and Abelson,
1977, p. 21.)
Just as AI must consider the same issues that are addressed in psychology
and linguistics, other aspects of knowledge, dealt with by other traditional
disciplines, must be considered. For example, current ideas in AI about
linking computing machines together into coherent systems or cooperative
problem-solvers forces us to consider the sociological aspects of knowing. A
fundamental problem in AI is communication among many individual units,
each of which knows some things relevant to some problems as well as
something about the other units. The form of communication between units,
the organizational structure of the complex, and the nature of the individ¬
uals’ knowledge of each other are all questions that must find some engineer¬
ing solution if the apparent power of distributed processing is to be realized.
These issues have been studied in other disciplines, albeit from very
different perspectives and with different goals and methods. We can view
the different control schemes proposed for interprocess communication, for
example, as attempts to design social systems of knowledgeable entities.
Our intuitions, once again, form the specifications for the first systems. Reid
G. Smith has proposed a contract net where individual entities negotiate
their role in attacking the problem via requests for assistance from other
processors, proposals for help in reply, and contracts indicating agreement
to delegate part of the problem to another processor; and W. Kornfeld and
Carl Hewitt have developed a model explicitly based on problem-solving in
the scientific community. [R. G. Smith, 1978; Kornfeld and Hewitt, 1981.]
Only after we have been able to build many systems based on such models
will we be able to identify key factors in the design of such systems.
There is another kind of study of the mind, conducted by scientists who
seek to understand the workings of the brain. The brain is a “mechanism’’
that has been associated with computing machines since their invention, and
that has puzzled computer scientists greatly:
We know the basic active organs of the nervous system (the nerve cells). There
is every reason to believe that a very large-capacity memory is associated with
this system. We do most emphatically not know what type of physical entities
are the basic components for the memory in question, (von Neumann, 1958,
p. 68.)
It might have been necessary a decade ago to argue for the commonality of the
information processes that are employed by such disparate systems as com¬
puters and human nervous systems. The evidence for that commonality is now
overwhelming, and the remaining questions about the boundaries of cognitive
science have more to do with whether there also exist nontrivial commonalities
with information processing in genetic systems than with whether men and
machines both think. (Simon, 1980, p. 45.)
Observing our own practice—that is, seeing what the computer implicitly tells
us about the nature of intelligence as we struggle to synthesize intelligent
systems—is a fundamental source of scientific knowledge for us. (Ibid., p. 19.)
Most actual designs have turned out initially to exhibit serious deficiencies, and
most predictions of performance have been startlingly inaccurate.
Under these circumstances, the main route open to the development and im¬
provement of time-sharing systems is to build them and see how they behave.
(Ibid., p. 21.)
Newell and Simon, in their view that computer programs are a vehicle for
expressing psychological theories rather than just serving as a metaphor for
mental processes, were already taking a strong position relative to even the
new breed of cognitive psychologists who were talking in terms of com¬
puterlike mental mechanisms. As Paul R. Cohen puts it, in his review of AI
work on models of cognition:
We should note that we have presented the strongest version of the informa¬
tion-processing approach, that advocated by Newell and Simon. Their position
is so strong that it defines information-processing psychology almost by exclu¬
sion: It is the field that uses methods alien to cognitive psychology to explore
questions alien to AI. This is an exaggeration, but it serves to illustrate why
there are thousands of cognitive psychologists, and hundreds of AI research¬
ers, and very few information-processing psychologists. (Cohen, 1982, p. 7.)
However, Newell and Simon did not stop there. A further development in
their thinking identified brains and computers as two species of physical
symbol system—the type of system which, they argue, must underlie any
intelligent behavior.
At the root of intelligence are symbols, with their denotative power and their
susceptibility to manipulation. And symbols can be manufactured of almost
anything that can be arranged and patterned and combined. Intelligence is
mind implemented by any patternable kind of matter. (Simon, 1980, p. 35.)
A physical symbol system has the necessary and sufficient means for general
intelligent action. (Newell and Simon, 1976a, p. 116.)
The basic point of view inhabiting our work has been that programmed com¬
puter and human problem solver are both species belonging to the genus IPS.
(Newell and Simon, 1972, p. 869.)
The question whether machines could think was certainly an issue in the
early days of AI research, although dismissed rather summarily by those
who shaped the emerging science:
To ask whether these computers can think is ambiguous. In the naive realistic
sense of the term, it is people who think, and not either brains or machines. If,
however, we permit ourselves the ellipsis of referring to the operations of the
brain as “thinking,” then, of course, our computers “think” .... (McCulloch,
1965, p. 368.)
Addressing fundamental issues like this one in their early writing, several
researchers suggested a parallel with the study of flight, considering cogni¬
tion as another natural phenomenon that could eventually be achieved by
machines.
Today, despite our ignorance, we can point to that biological milestone, the
thinking brain, in the same spirit as the scientists many hundreds of years ago
pointed to the bird as a demonstration in nature that mechanisms heavier than
air could fly. (Feigenbaum and Feldman, 1963, p. 8.)
Consider how people came to understand how birds fly. Certainly we observed
birds. But mainly to recognize certain phenomena. Real understanding of bird
flight came from understanding flight; not birds. (Papert, 1972, pp. 1-2.)
Every new design brings new data about what works, what does not, and
clues as to why. Every new contraption tries some different design alterna¬
tive in the space defined by our theory-language. And every attempt clarifies
our understanding of what it means to fly.
But there is more to the sciences of the artificial than defining the true
nature of natural phenomena. The exploration of the artifacts themselves,
the stiff-winged flying machines, because they are useful to society, will
naturally extend the exploration of various points of interface between the
technology and society. While nature’s exploration of the possibilities is
limited by its mutation mechanism, human inventors will vary every param¬
eter they can think of to produce effects that might be useful—exploring
constraints on the design of their machines from every angle. The space of
“flight” phenomena will be populated by examples that nature has not had a
chance to try.
This argument, that the utility of intelligent machines will drive the explora¬
tion of their capabilities, suggests that the development of AI technology has
begun an exploration of cognitive phenomena that will involve aspects of
cognition that are not easy to study in nature. In fact, as with the study of
flight, AI will enable us to see natural intelligence as a limited capability, in
terms of design tradeoffs made in the evolution of biological cognition.
We build computers and programs for many reasons. We build them to serve
society and as tools for carrying out the economic tasks of society. But as basic
scientists we build machines and programs as a way of discovering new phe¬
nomena and analyzing phenomena we already know about. . . . the phenomena
surrounding computers are deep and obscure, requiring much experimentation
to assess their nature. (Newell and Simon, 1976u, p. 114.)
For what will AI systems be useful? How will they be involved in the
economic tasks of society? It has certainly been argued that this point is one
that distinguishes biological systems from machines:
The human is a physical symbol system, yes, with a component of pure cogni¬
tion describable by mechanisms. . . . But the human is more: The human is an
animate organism, with a biological basis and an evolutionary and cultural
history. Moreover, the human is a social animal, interacting with others, with
the environment, and with itself. The core disciplines of cognitive science have
tended to ignore these aspects of behavior .... (Norman, 1980, pp. 3-4.)
Al: COGNITION AS COMPUTATION 257
The difference between natural and artificial devices is not simply that they are
constructed of different stuff; their basic functions differ. Humans survive ....
(Ibid., p. 10.)
Tools evolve and survive based on their utility to the people who use them.
Either the users find better tools, or their competitors find them. This pro¬
cess will certainly continue with the development of cognitive tools and will
dramatically change the way we think about AI.
In fact, this change has already begun in AI laboratories, but the place where
the changing perception of AI systems is most dramatic and accelerated is,
not surprisingly in our society, the marketplace.
Al, INC.
Expert Systems
One needs to move toward task environments of greater complexity and open¬
ness—to everyday reasoning, to scientific discovery, and so on. The tasks we
tackled, though highly complex by prior psychological standards, still are sim¬
ple in many respects. (Newell and Simon, 1972, p. 872.)
Humans have difficulty keeping track of all the knowledge that might be
relevant to a problem, exploring all alternative solution-paths, and making
sure none of the valid solutions is overlooked in the process. Work on
DENDRAL showed that when the human expert could explain exactly what
he or she was doing in solving problems, the machine could achieve expert-
level performance.
Continued research at Stanford’s Heuristic Programming Project next
produced the MYCIN system, an experiment in modeling medical diagnostic
reasoning. [Shortliffe, 1976.] In production rules of the form if (condition)
then (action), Shortliffe encoded the kind of information about their rea¬
soning process that physicians were most able to give—advice about what to
do in certain situations. In other words, the if parts of the rules contain
clauses that attempt to differentiate a certain situation, and the then part
describes what to do if we find ourselves in that situation. This production-rule
knowledge representation worked surprisingly well; MYCIN was able to
perform its task in a specific area of infectious-disease diagnosis as well as
the best experts in the country.
Furthermore, the MYCIN structure was seen to be, at least to some
extent, independent of the domain of medicine. So long as experts could
describe their knowledge in terms of if-then rules, the reasoning mechanism
that MYCIN used to make inferences from a large set of rules would come
up with the right questions and eventually, a satisfactory analysis. MYCIN-
like systems have been successfully built in research laboratories for appli¬
cations as diverse as mineral exploration, diagnosis of computer-equipment
failure, and even for advising users about how to use complex systems.
Transfer of Expertise
A key idea in our current approach to building expert systems is that these
programs should not only be able to apply the corpus of expert knowledge to
Al: COGNITION AS COMPUTATION 259
specific problems, but they should also be able to interact with the users and
experts just as humans do when they learn, explain, and teach what they know.
. . . these transfer of expertise (TOE) capabilities were originally necessitated
by “human engineering” considerations—the people who build and use our
systems needed a variety of “assistance” and “explanation” facilities. How¬
ever, there is more to the idea of TOE than the implementation of needed user
features: These social interactions—learning from experts, explaining one’s
reasoning, and teaching what one knows—are essential dimensions of human
knowledge. They are as fundamental to the nature of intelligence as expert-
level problem-solving, and they have changed our ideas about representation
and about knowledge. (Barr, Bennett, and Clancey, 1979, p. 1.)
We are building systems that take part in the human activity of transfer of
expertise among experts, practitioners, and students in different kinds of do¬
mains. Our problems remain the same as they were before: We must find good
ways to represent knowledge and meta-knowledge, to carry on a dialogue, and
to solve problems in the domain. But the guiding principles of our approach
and the underlying constraints on our solutions have subtly shifted: Our sys¬
tems are no longer being designed solely to be expert problem solvers, using
vast amounts of encoded knowledge. There are aspects of “knowing” that
have so far remained unexplored in AI research: by participation in human
transfer of expertise, these systems will involve more of the fabric of behavior
that is the reason we ascribe knowledge and intelligence to people. (Barr,
Bennett, and Clancey, 1979, p. 5.)
With these systems, it will at last become economical to match human beings in
real time with really large machines. This means that we can work toward
programming what will be, in effect, “thinking aids.” In the years to come we
expect that these man-machine systems will share, and perhaps for a time be
dominant, in our advance toward the development of “artificial intelligence.”
(Minsky, 1963, p. 450.)
Among the scientific disciplines that study knowledge, the potential for com¬
mercial applications of artificial intelligence presents unique opportunities.
In order to identify and fill the niches in which intelligent machines will
survive, we must ask questions about knowledge from a rather different
perspective. We must identify the role that various aspects of intelligence
play, or could play, in the affairs of people, in such a way that we can
identify correctable shortcomings in the way things are done.
There is no question that the current best design of an intelligent system,
the human brain, has its limitations. Computers have already been used to
aid people in dealing with such shortcomings as memory failure and confu¬
sions, overloading in busy situations, their tendency to boredom, and their
need for sleep. These extended capabilities—total recall, rapid processing,
and uninterrupted attention—are cognitive capabilities that we have been
willing to concede to the new species in genus symbol manipulators. They
have helped us do the things we did before and have made some entirely new
capabilities possible; for example, airline reservation systems, 24-hour
banking, and Pac-Man (although the truly challenging computer games are
yet to come!). Intelligence is also going to be present in this new species, as
envisioned twenty years ago by Marvin Minsky: “I believe . . . that we are
on the threshold of an era that will be strongly influenced, and quite possibly
dominated, by intelligent problem-solving machines.” (Minsky, 1963, p.
406.) Finding a way to apply this new intellectual capability, for effectively
applying relevant experience to new situations, is the task ahead for AI, Inc.
It may be a while in coming, and it may involve a rethinking of the way we go
about some cognitive activities. But it is extremely important that the devel¬
opment of intelligent machines be pursued, for the human mind is not only
limited in its storage and processing capacity, it also has known bugs: It can
easily be misled, stubborn, and even blind to the truth, especially when
pushed to the limits.
And, as is nature’s way, everything gets pushed to the limits, including
humans. We must find a way of organizing ourselves more effectively, of
bringing together the energies of larger groups of people toward a common
goal. Intelligent systems, built from computer and communications technol¬
ogy, will someday know more than any individual human about what is
going on in complex enterprises involving millions of people, like a multina¬
tional corporation or a city. And intelligent systems will be able to explain
262 AVRON BARR
each person’s part of the task. We will build more productive factories this
way and maybe someday, a more peaceful world. We must keep in mind,
following our analogy with flight, that the capabilities of intelligence as it
exists in nature are not necessarily its natural limits: “There are other facets
to this analogy with flight; it, too, is a continuum, and some once thought
that the speed of sound represented a boundary beyond which flight was
impossible.” (Armer, 1963, p. 398.)
ARTIFICIAL INTELLIGENCE
Subcognition as Computation
Douglas R. Hofstadter
The philosopher John Searle has recently made quite a stir in the cognitive-
science and philosophy-of-mind circles with his celebrated “Chinese room”
thought experiment, whose purpose is to reveal as illusory the aims of
artificial intelligence (AI), and particularly to discredit what he labels
“strong AI”—the belief that a programmed computer can, in principle, be
conscious. [Searle, 1981.] Various synonymous phrases could be substituted
for “be conscious” here, such as: “think,” “have a soul” (in a humanistic
rather than a religious sense), “have an inner life,” “have semantics” (as
distinguished from “mere syntax”), “have content” (as distinguished from
“mere form”), “have intentionality,” “be something it is like something to
be” (a somewhat ponderous yet appealing phrase due to philosopher
Thomas Nagel), “have personhood,” and others. Each of these phrases has
its own peculiar set of connotations and imagery attached to it as well as its
own history and proponents. For our purposes, however, we shall consider
them all as equivalent and lump them all together, so that the claim of strong
AI now becomes very strong indeed.
At the same time, various AI workers have been developing their own
philosophies of what AI is and have developed some useful terms and slo¬
gans to describe their endeavor. Some of them are: information processing,
cognition as computation, physical symbol system, symbol manipulation,
expert system, and knowledge engineering. There is some confusion as to
what words like symbol and cognition actually mean, just as there is some
confusion as to what words like semantics and syntax mean.
It is the purpose of this paper to try to delve into the meanings of such
elusive terms and, at the same time, shed some light on the views of John
Searle, on the one hand, and Allen Newell and Herbert Simon, on the other
hand—visible AI pioneers who are responsible for several of the terms in the
previous paragraph. The thoughts expressed herein were originally triggered
I would like to thank Daniel C. Dennett and Marsha J. Meredith for their careful reading of, and
valuable comments on, earlier drafts of this paper.
263
264 DOUGLAS R. HOFSTADTER
itself. Here, that “thing” is a very abstract concept—namely, the way that
this particular shape manifests the abstract quality of being an a. The prob¬
lem of a is, thus, intimately connected with the problems of b through z and
with that of stylistic consistency.
The existence of optical character readers might lead us to believe at first
that the letter-recognition problem has been solved. If we consider the prob¬
lem a little more carefully, however, we see that the surface has barely been
scratched. In truth, the way that most optical character-recognition pro¬
grams work is by a fancy kind of template matching, in which statistics are
done to determine which character, out of a fixed repertoire of, say, 100
stored characters, is the “best match.” This is about like assuming that the
way I recognize my mother is by comparing the scene in front of me with
stored memories of the appearances of tigers, cigarettes, hula hoops, gam¬
bling casinos, and can openers (and of course all other things in the world
simultaneously) and somehow instantly coming up with the “best match.”
The problem of recognizing letters of the alphabet is no less deep than that of
recognizing your mother, even if it might seem so, given that the number of
Platonic prototype items is on the small side (26, if we ignore all characters
but the lowercase alphabet). We can even narrow it down further—to just a
handful. As a matter of fact, Godfried Toussaint, editor of the pattern-
recognition papers for the IEEE Transactions, has said to me that he would
like to put up a prize for the first program that could say correctly, of 20
characters that people easily can identify, which are a's and which are b's.
To carry out such a task, a program cannot just recognize that a shape is an
a; it has to see how that shape embodies n-ness. And then, as a test of
whether the program really knows its letters, it would have to carry “that
style” over to the other letters of the alphabet. This is the goal of my
research: To find out how to make letters slip in “similar ways to each
other,” so as to constitute a consistent artistic style in a typeface—or simply
a consistent way of writing the alphabet.
By contrast, most AI work on vision pertains to such things as aerial
reconnaissance or robot-guidance programs. This would suggest that the
basic problem of vision is to figure out how to recognize textures and how to
mediate between two and three dimensions. But what about the fact that
although we are all marvelous face-recognizers, practically none of us can
draw a face at all well—even of someone we love? Most of us are flops at
drawing even such simple things as pencils and hands and books. I person¬
ally have learned to recognize hundreds of Chinese characters (shapes that
involve neither three dimensions nor textures) and yet, on trying to repro¬
duce them from memory, find myself often drawing confused mixtures of
Al: SUBCOGNITION AS COMPUTATION 267
characters, leaving out basic components, or, worst of all, being unable to
recall anything but the vaguest “feel” of the character and not being able to
draw a single line.
Closer to home, most of us have read literally millions of, say, k's with
serifs, yet practically none of us can draw a k with serifs in the standard
places. (This holds, of course, for any letter of the alphabet.) I suspect that
many people—perhaps most—are not even consciously aware of the fact
that there are two different types of lowercase a and of lowercase g, just as
many people seem to have a very hard time drawing a distinction between
lowercase and uppercase letters, and a few have a hard time telling letters
drawn forwards from letters drawn backwards.
How can such a fantastic “recognition machine” as our brain be so ter¬
rible at rendition? Clearly, there must be something very complex going on,
enabling us to accept things as members’ categories and to perceive how
they are members of those categories, yet not enabling us to reproduce those
things from memory. This is a deep mystery.
In his book Pattern Recognition, Mikhail Bongard concludes with a series
of 100 puzzles for a visual pattern-recognizer, whether human, machine, or
alien, and, to my mind, it is no accident that he caps off his set with letter-
forms. [Bongard, 1970.] In other words, he works his way up to letterforms
as being at the pinnacle of visual recognition ability. There exists no pattern-
recognition program in the world today that can come anywhere close to
doing those Bongard problems. [Ibid.] And yet Barr cites Simon as writing
the following statement:
The evidence for that commonality [between the information processes that
are employed by such disparate systems as computers and human nervous
systems] is now overwhelming, and the remaining questions about the bound¬
aries of cognitive science have more to do with whether there also exist non¬
trivial commonalities with information processing in genetic systems than
with whether men and machines both think. (Simon, 1980, p. 45.)
I find it difficult to understand how Simon can believe this in an era when
computers still cannot do basic kinds of subcognitive acts (acts that we feel
are unconscious, acts that underlie cognition).
In another lecture in 1979 (the opening lecture of the first meeting of the
Cognitive Science Society, also in San Diego), I recall Simon proclaiming
that despite much doubting by people not in the know, there is no longer any
question as to whether computers can think. If he had meant that there
should no longer be any question about whether machines may eventually
become able to think or whether we humans are machines (in some abstract
sense of the term), then I would be in accord with his statement. But after
hearing and reading such statements over and over again, I don’t think that s
what he meant at all. I get the impression that Simon genuinely believes that
today’s machines are intelligent and that they really do think (or perform
268 DOUGLAS R. HOFSTADTER
“acts of cognition”—to use a bit of jargon that adds nothing to the meaning
but makes it sound more scientific). I will come back to that shortly, since it
is in essence the central bone of contention in this article but first, a few
more remarks on AI domains.
tions. The domain was phenomenally far from being understood by a single,
unified program. Here, then, was a nearly ideal domain for exploring what
cognition truly is—and it was suddenly dropped. Researchers at Massachu¬
setts Institute of Technology were at one time doing truly basic research on
intelligence and then quit. Much basic research has been supplanted by large
teams marketing what they vaunt as “knowledge engineering.” Firmly
grounded engineering is fine, but it seems to me that this type of engineering
is not built on the solid foundations of a science, but on a number of recipes
that have worked with some success in limited domains.
In my opinion, the proper choice of domain is the critical decision that an
AI researcher makes when beginning a project. If you choose to get involved
in medical diagnosis at the expert level, then you are going to get mired
down in a host of technical problems that have nothing to do with how the
mind works. The same goes for the other earlier-cited ponderous domains
that current work in expert systems involves. By contrast, if you are in
control of your own domain, and can tailor it and prune it so that you keep
the essence of the problem while getting rid of extraneous features, then you
stand a chance of discovering something fundamental.
Early programs on the nature of analogy [Evans, 1968], sequence ex¬
trapolation [Simon and Kotovsky, 1963, among others], and so on, were
moving in the right direction. But then, somehow, it became a common
notion that these problems had been solved. Simply because Thomas Evans
had made a program that could do some very restricted types of visual-
analogy problem “as well as a high school student,” many people thought
the book was closed. However, we need only look at Bongard’s 100 to see
how hopelessly far we are from dealing with analogies. [Bongard, 1970.] We
need only look at any collection of typefaces (look at any magazine’s adver¬
tisements for a vast variety) to see how enormously far we are from under¬
standing letterforms. As I claimed earlier, letterforms are probably the quint¬
essential problem of pattern recognition. It is both baffling and disturbing to
me to see so many people working on imitating cognitive functions at the
highest level of sophistication when their programs cannot carry out cogni¬
tive functions at much lower levels of sophistication.
There are some notable exceptions. The Schank group at Yale, whose origi¬
nal goal was to develop a program that could understand natural language,
has been forced to retreat, and to devote most of its attention to the organi¬
zation of memory, which is certainly at the crux of cognition (because it is
part of subcognition, incidentally)—and the group has gracefully accom¬
modated this shift of focus. [Schank, 1982.] I will not be at all surprised,
however, if eventually the group is forced into yet further retreats—in fact,
all the way back to Bongard problems or the like. Why? Simply because
270 DOUGLAS R. HOFSTADTER
their work (on such things as how to discover what “adage” accurately
captures the “essence” of a story or episode) already has led them into the
deep waters of abstraction, perception, and classification. These are the
issues that Bongard problems illustrate so perfectly. Bongard problems are
the idealized (“frictionless”) versions of these critical questions.
It is interesting that Bongard problems are in actuality nothing other than
a well-worked-out set of typical IQ-test problems, the kind that Lewis Ter-
man and Alfred Binet first invented 60 or more years ago. [Terman, 1916;
Binet, 1916.] Over the years, many other less talented people have invented
similar visual puzzles that had the unfortunate property of being filled with
ambiguity and multiple answers. This (among other things) has given IQ
tests a bad name. Whether or not IQ is a valid concept, however, there can
be little question that the original insight of Terman and Binet—that care¬
fully constructed simple visual-analogy problems probe close to the core
mechanisms of intelligence—is correct. Perhaps the political climate created
a kind of knee-jerk reflex in many cognitive scientists, causing them to shy
away from anything that smacked of IQ tests, since issues of cultural bias
and racism began raising their ugly heads. But we need not be so Pavlovian
as to jump whenever a visual-analogy problem is placed in front of us. In
any case, it will be good when AI people are finally driven back to looking
at the insights of people working in the 1920s and 1930s, such as Wittgenstein
and his “games,” Kohler and Koffka and Wertheimer and their “gestalts,”
and Terman and Binet and their IQ-test problems. [Ibid.; Wittgenstein, 1922;
Wertheimer, 1925; Kohler, 1929; Koffka, 1935.]
I was saying that some AI groups seem to be less afraid of toy domains,
or, more accurately put, they seem to be less afraid of stripping down their
domain in successive steps, to isolate the core issues of intelligence that it
involves. Aside from the Schank group, N. Sridharan at Rutgers University
has been doing some very interesting work on prototype deformation,
which, although it springs from work in legal reasoning in a quite messy real-
world domain, has been abstracted into a form in which it is perhaps more
like a toy domain (or, perhaps less pejorative-sounding, an idealized domain)
than at first would appear. [Sridharan, 1980.] The Lindsay-Norman-
Rumelhart group at San Diego has been doing work for years on understand¬
ing errors, such as grammatical slips, typing errors, errors in everyday phys¬
ical actions (such as winding your watch when you mean to switch television
channels), for the insights it may offer into the underlying (subcognitive)
mechanisms. [Norman, 1981.]
Then there are those people who are working on various programs for
perception, whether visual or auditory. One of the most interesting was
Hearsay-II, a speech-understanding program developed at Carnegie-Mellon
University, Simon’s home. It is, therefore, very surprising to me that Simon,
who surely was very aware of the wonderfully intricate and quite beautiful
architecture of Hearsay-II, could then make a comment indicating that per¬
ception and, in general, subcognitive (under 100 milliseconds) processes,
have no interest. [Reddy, 1976.]
AI: SUBCOGNITION AS COMPUTATION 271
There are surely many other less publicized groups that are also working
on more humble domains and on more pure problems of mind, but from
looking at the proceedings of AI conferences, we might get the impression
that, indeed, computers must really be able to think these days, since, after
all, they are doing anything and everything cognitive—from opthalmology to
biology to chemistry to mathematics—even discovering scientific laws from
looking at tables of numerical data, to mention one project (“Bacon”) that
Simon has been involved in.
The problem is that AI programs are carrying out all these cognitive activi¬
ties in the absence of any subcognitive activity. There is no substrate that
corresponds to what goes on in the brain. There is no fluid recognition and
recall and reminding. These programs have no common sense, little sense of
similarity or repetition or pattern. They can perceive some patterns as long
as they have been anticipated—and, particularly, as long as the place where
they will occur has been anticipated—but they cannot see patterns where
nobody told them explicitly to look. They do not learn at a high level of
abstraction.
This style is in complete contrast to how people are. People perceive
patterns anywhere and everywhere, without knowing in advance where to
look. People learn automatically in all aspects of life. These are just facets of
common sense. Common sense is not an area of expertise, but a general
that is, domain-independent—capacity that has to do with fluidity in repre¬
sentation of concepts, an ability to sift what is important from what is not, an
ability to find unanticipated analogical similarities between totally different
concepts (reminding, as Schank calls it). We have a long way to go before
our programs exhibit this cognitive style.
Recognition of one’s mother’s face is still nearly as much of a mystery as
it was 30 years ago. And what about such things as recognizing family
resemblances between people, recognizing a French face, recognizing kind¬
ness or earnestness or slyness or harshness in a face? Even recognizing
age—even sex!—these are fantastically difficult problems! As Donald
Knuth has pointed out, we have written programs that can do wonderfully
well at what people have to work very hard at doing consciously (e.g., doing
integrals, playing chess, medical diagnosis, etc.)—but we have yet to write a
program that remotely approaches our ability to do what we do without
thinking or training—things like understanding a conversation partner with
an accent at a loud cocktail party with music blaring in the background,
while at the same time overhearing wisps of conversations in the far corner
of the room. [Knuth, 1974a.] Or perhaps finding our way through a forest on
an overgrown trail. Or perhaps just doing some anagrams absentmindedly
while washing the dishes.
Asking for a program that can discover new scientific laws without having
272 DOUGLAS R. HOFSTADTER
Well, I feel we’re much further away from human-level performance than
Simon does. I, for one, would like to see a program that does anagrams the
way a person does. Why anagrams? Because they are a toy domain where
some very significant subcognitive processes play the central role.
What I mean is this. When we look at a “jumble” such as telkin in the
newspaper, we immediately begin shifting around letters into tentative
groups, making such stabs as knitle, klinte, linket, keltin, tinkle—and then
we notice that, indeed, tinkle is a word. The part of this process that I am
interested in is the part that precedes the recognition of tinkle as a word. It’s
that part that involves experimentation, based only on the style or “feel” of
English words—using intuitions about letter affinities, plausible clusters and
their stabilities, syllable qualities, and so on. When we first read a jumble in
the newspaper, we play around, rearranging, regrouping, reshuffling, in
complex ways that we have no control over. In fact, it feels as if we throw
the letters up into the air separately, and when they come down, they have
somehow magically “glommed” together in some English-like word! It’s a
marvelous feeling—and it is anything but cognitive, anything but conscious.
(Yet, interestingly, we take credit for being good at anagrams, if we are
good!)
It turns out that most literate people can handle “jumbles” (my term for
single-word anagrams) of five or six letters, sometimes seven or eight letters;
with practice, maybe even ten or twelve. But beyond that, it gets very hard
to keep the letters in our head. It is especially hard if there are repeated
letters, since we tend to get confused about which letters there are multiple
copies of. (In one case, I rearranged the letters dinnal into nadlid—
incorrectly. You can try raregarden if you dare.) Now in one sense, the fact
that the problem gets harder and harder with more and more letters is hardly
surprising. It is obviously related to the famous “7 plus or minus 2” figure
that psychologist George A. Miller first reported in connection with short¬
term memory capacity. [Miller, 1956.] But there are different ways of inter¬
preting such a connection.
One way to think that this might come about is to assume that concepts
for the individual letters get ‘activated” and then interact. When too many
Al: SUBCOGNITION AS COMPUTATION 273
AI people have not explored. How do those letters swirl among one another,
fluidly and tentatively making and breaking alliances? “Glomming” to¬
gether, then coming apart, almost like little biological objects in a cell. AI
people have not paid much attention to such problems as anagrams. Perhaps
they would say that the problem is already solved. After all, a virtuoso
programmer has made a program print out all possible words that anagram¬
matize into other words in English. Or perhaps they would point out that in
principle, you can do an “alphabetize” followed by a “hash” and thereby
retrieve from any given set of letters all the words they anagrammatize into.
Well, this is all fine and dandy, but it is really beside the point. It is merely a
show of brute force and has nothing to contribute to our understanding of
how we actually do anagrams ourselves, just as most chess programs have
absolutely nothing to say about how chess masters play (as Adriaan de
Groot, and later, William Chase and Herbert Simon have pointed out), [de
Groot, 1965; Chase and Simon, 1973.]
Is anagrams simply a trivial, silly, “toy” domain? Or is it serious? I
maintain that it is a far purer, far more interesting domain than many of the
complex real-world domains of the expert systems, precisely because it is so
playful, so unconscious, so enjoyable for people. It is obviously more re¬
lated to creativity and spontaneity than it is to logical derivations, but that
does not make it—or the mode of thinking that it represents—any less
worthy of attention. In fact, because it epitomizes the unconscious mode of
thought, I think it more worthy of attention.
In short, it seems to me that something fundamental is missing in the
orthodox AI “information-processing” model of cognition, and that is some
sort of substrate from which intelligence emerges as an epiphenomenon.
Most AI people do not want to tackle that kind of underpinning work. Could
it be that they really believe that machines already can think, already have
concepts, already can do analogies? It seems that a large camp of AI people
really do believe these things.
top level need not have anything to do with the underlying swarm of activi¬
ties bringing it into existence. In particular, something can be computational
at one level but not at another level.
Yet, many AI people, despite considerable sophistication in thinking
about a given system at different levels, still seem to miss this. Most AI work
goes into efforts to build rational thought (cognition) out of smaller rational
thoughts (elementary steps of deduction, for instance, or elementary mo¬
tions in a tree). It comes down to thinking that what we see at the top level of
our minds—our ability to think—comes out of rational “information¬
processing” activity, with no deeper levels below that.
Many interesting ideas, in fact, have been inspired by this hope. I find
much of the work in AI to be fascinating and provocative, yet somehow I
feel dissatisfied with the overall trend. For instance, there are some people
who believe that the ultimate solution to AI lies in getting better and better
theorem-proving mechanisms in some predicate calculus. They have devel¬
oped extremely efficient and novel ways of thinking about logic. Some peo¬
ple—Simon and Newell, particularly—have argued that the ultimate solu¬
tion lies in getting more and more efficient ways of searching a vast space of
possibilities. (They refer to selective heuristic search as the key mechanism
of intelligence.) Again, many interesting discoveries have come out of this.
Then there are others who think that the key to thought involves making
some complex language in which pattern-matching or backtracking or inher¬
itance or planning or reflective logic is easily carried out. Now admittedly,
such systems, when developed, are good for solving a large class of prob¬
lems, exemplified by such AI chestnuts as the missionary-and-cannibals
problem, “cryptarithmetic” problems, retrograde chess problems, and
many other specialized sorts of basically logical analysis. However, these
kinds of techniques of building small logical components up to make large
logical structures have not proven good for such things as recognizing your
mother or drawing the alphabet in a novel and pleasing way.
One group of AI people who seem to have a different attitude consists of
those who are working on problems of perception and recognition. There,
the idea of coordinating many parallel processes is important, as is the idea
that pieces of evidence can add up in a self-reinforcing way, so as to bring
about the locking-in of a hypothesis that no one of the pieces of evidence
could on its own justify. It is not easy to describe the flavor of this kind ot
program architecture without going into multiple technical details. How¬
ever, it is very different in flavor from ones operating in a world where
everything comes clean and precategorized—where everything is specified
in advance: “There are three missionaries and three cannibals and one boat
and one river and . . . ,” which is immediately turned into a predicate-
calculus statement or a frame representation, ready to be manipulated by an
“inference engine.” The missing link seems to be the one between percep¬
tion and cognition, which I would rephrase as the link between subcognition
and cognition, that gap between the sub-100-millisecond world and the
super-100-millisecond world.
276 DOUGLAS R. HOFSTADTER
One way to explore this disagreement is to look at some of the ways that
Simon and Newell express themselves about symbols.
At the root of intelligence are symbols, with their denotative power and their
susceptibility to manipulation. And symbols can be manufactured of almost
anything that can be arranged and patterned and combined. Intelligence is
mind implemented by any patternable kind of matter. (Simon, 1980, p. 35.)
From this quotation and from others, we can see that to Simon and Newell, a
symbol seems to be any token, any character inside a computer that has an
ASCII code (a standard but arbitrarily assigned sequence of seven bits). To
me, by contrast, symbol connotes something with representational power.
To them (if I am not mistaken), it would be fine to call a bit (inside a
computer) or a neuron-firing a symbol; however, I cannot feel comfortable
with that usage of the term.
To me, the crux of the word symbol is its connection with the verb to
symbolize, which means to denote, to represent, to stand for, and so on.
Now in the preceding quotation, Simon refers to the denotative power of
symbols—yet, in another part of his paper, Barr quotes Newell and Simon
as saying that thought is “the manipulation of formal tokens.” [Newell and
Simon, 1976a.] It is not clear to me which side of the fence they really are
on.
It takes an immense amount of richness for something to represent some¬
thing else. The letter I does not in and of itself stand for the person I am or
for the concept of selfhood. That quality comes to it from the way that the
word behaves in the totality of the English language. It comes from a mas¬
sively complex set of usages and patterns and regularities, ones that are
regular enough for babies to be able to detect, so that they, too, eventually
come to say “I” to talk about themselves.
Formal tokens such as I or hamburger are in themselves empty; they do
not denote. Nor can they be made to denote in the full rich intuitive sense of
Al: SUBCOGNITION AS COMPUTATION 111
the term by having them obey some rules. You can’t simply push around
some Pnames of LISP atoms according to complex rules and hope to come
out with genuine thought or understanding. (This, by the way, is probably a
charitable way to interpret Searle’s point in his previously mentioned pa¬
per—namely, as a rebellion against claims that programs that can manipu¬
late tokens such as John, ate, a, hamburger actually have understanding.)
Manipulation of empty tokens is not enough to create understanding—
although it is enough to imbue them with meaning in a limited sense of the
term, as I stress repeatedly in my book Godel, Escher, Bach: an Eternal
Golden Braid. [Hofstadter, 1979, particularly chaps. 2-6.]
It is the vast collections of ants (read neural firings, if you prefer) that add up
to something genuinely symbolic. And who can say whether there exist
rules—formal, computational rules—at the level of the teams themselves
(read concepts, ideas, thoughts) that are of full predictive power in describ¬
ing how they will flow? I am speaking of rules that allow us to ignore what is
going on “down below” yet still yield perfect or at least very accurate
predictions of the teams’ behavior.
To be sure, there are phenomenological observations that can be for¬
malized to sound like rules that will describe, very vaguely, how those
highest-level teams act. But what guarantee is there that we can skim off the
full fluidity of the top-level activity of a brain, and encapsulate it—without
any lower substrate—in the form of some computational rules?
To ask an analogous question, what guarantee is there that there are rules
at the cloud level (more properly speaking, the level of cold fronts, isobars,
trade winds, etc.) that will allow us to say accurately how the atmosphere is
going to behave on a large scale? Perhaps there are no such rules; perhaps
weather prediction is an intrinsically intractable problem. Perhaps the be¬
havior of clouds is not expressible in terms that are computational at then-
own level, even if the behavior of the microscopic substrate—the
molecules—is computational.
The premise of AI is that thoughts themselves are computational entities
at their own level. At least, this is the premise of the information-processing
school of AI, and I have very serious doubts about it.
The difference between my active symbols (teams) and passive informa¬
tion-processing symbols (ants, tokens) is that the active symbols flow and
act on their own. In other words, there is no higher-level agent (read pro¬
gram) that reaches down and shoves them around. Active symbols must
incorporate within their own structures the wherewithal to trigger and cause
actions. They cannot just be passive storehouses, bins, or receptacles of
data. Yet to Newell and Simon, it seems, even so tiny a thing as a bit is a
symbol. This is brought out repeatedly in their writings on physical symbol
systems.
Al: SUBCOGNITION AS COMPUTATION 279
A good term for the little units that a computer manipulates (as well as for
neuron firings) is tokens. All computers are good at token manipulation;
however, only some—the appropriately programmed ones—could support
active symbols. (I prefer not to say that they would carry out symbol manip¬
ulation, since that gets back to that image of a central program shoving
around some passive representational structures.) The point is, in such a
hypothetical program (and none exists as yet), the symbols themselves are
acting!
A simple analogy from ordinary programming might help to convey the
level distinction that I am trying to make here. When a computer is running a
LISP program, does it do function calling? To say yes would be unconven¬
tional. The conventional intuition is that functions call other functions and
the computer is simply the hardware that supports function-calling activity.
In somewhat the same sense, although with much more parallelism, symbols
activate or trigger or awaken other symbols in a brain.
The brain itself does not manipulate symbols; the brain is the medium in
which the symbols are floating and in which they trigger each other. There is
no central manipulator, no central program. There is simply a vast collection
of “teams”—patterns of neural firings that, like teams of ants, trigger other
patterns of neural firings. The symbols are not “down there ’ at the level of
the individual firings; they are “up here” where we do our verbalization. We
feel those symbols churning within ourselves in somewhat the same way we
feel our stomach churning. We do not do symbol manipulation by some sort
of act of will, leave alone some set of logical rules of deduction. We cannot
decide what we will next think of nor how our thoughts will progress.
Not only are we not symbol manipulators; in fact, quite to the contrary,
we are manipulated by our symbols! As Scott Kim has put it, rather than
speak of “free will,” perhaps it is more appropriate to speak of “free
won’t.” This way of looking at things turns everything on its head, placing
cognition—that rational-seeming level of our minds—where it belongs,
namely, as a consequence of much deeper processes of myriads of interact¬
ing subcognitive structures. The rational has had entirely too much made of
it in AI research; it is time for some of the irrational and subcognitive to be
recognized (no pun intended) for its pivotal role.
genuinely have tried to add the two numbers at the symbol level and made a
mistake, just as you or I might have, despite having neurons that can add
fast.
The point is simply that the lower-level arithmetical processes (the adds,
the shifts, the multiplies, etc.) out of which the higher level of any AI pro¬
gram is composed are completely shielded from its view. To be sure,
Winograd could have artificially allowed his program to write little pieces of
LISP code that would execute and return answers to questions in English
such as “What is 720 factorial?” but that would be similar to our trying to
take advantage of the fact that we have billions of small analog adders in our
brain some time when we are trying to check a long grocery bill. We simply
don’t have access to those adders! We can’t reach them.
Let me, for a final time, make clear how this is completely in contradistinc¬
tion to standard computer programs. In a normal program, you can account
for every single operation at the bit level, by looking “upwards” toward the
top-level program. You can trace a high-level function call downwards: It
calls subroutines that call other subroutines that call this particular machine-
language routine that uses these words and in which this particular bit lies.
So there is a high-level, global reason why this particular bit is being manipu¬
lated.
By contrast, in an ant colony, a particular ant’s foray does not achieve
some global purpose. It has no interpretation in terms of the overall colony’s
goals; only when many such actions are considered at once does their statis¬
tical quality then emerge as purposeful, or interpretable. Ant actions are not
the translation into machine language of some “colony-level” program. No
one ant is essential; even large numbers of ants are dispensable. All that
matters are the statistics: Thanks to them the information moves around at a
level far above that of the ants; ditto for neural firings in brains. Not ditto for
most current AI programs’ architecture.
AI researchers started out thinking that they could reproduce all of cogni¬
tion through a 100 per cent top-down approach: functions calling subfunc¬
tions calling subsubfunctions, and so forth, until it all bottomed out in some
primitives. Thus, intelligence was thought to be hierarchically decompos¬
able, with cognition at the top driving subcognition at the bottom. There
were some successes and some difficulties—difficulties particularly in the
realm of perception. 1 hen along came such things as production systems and
pattern-directed inference. Here, some bottom-up processing was allowed
Al: SUBCOGNITION AS COMPUTATION 285
to occur within essentially still a top-down context. Gradually, the trend has
been shifting, but there still is a large element of top-down quality in AI.
It is my belief that until AI has been stood on its head and is 100 per cent
bottom-up, it won’t achieve the same level or type of intelligence as humans
have. To be sure, when that kind of architecture exists, there will still be
high-level, global, cognitive events—but they will be epiphenomenal, like
those in a brain. They will not in themselves be computational. Rather, they
will be constituted out of, and driven by, many many smaller computational
events rather than the reverse. In other words, subcognition at the bottom
will drive cognition at the top. And, perhaps most importantly, the activities
that take place at that cognitive top level will neither have been written nor
anticipated by any programmer. [Hofstadter, 1982/?; 1982d.]
Let me then close with a return to the comment of Simon’s: “Nothing
below 100 milliseconds is of interest in the study of cognition.’’ [Simon,
1981.] I cannot imagine a remark about AI with which I could more vehe¬
mently disagree. Simon seems to be most concerned with having programs
that can imitate chains of serial actions that come from verbal protocols of
various experimental subjects. Perhaps, in some domains, even in some
relatively complex and technical ones, people have come up with programs
that can do this. But what about the simpler, noncognitive acts that in reality
form the substrate for those cognitive acts? Whose program carries those out?
At present, no one’s. Why is this?
It is because AI people have, in general, tended to cling to a notion that in
some sense, thoughts obey formal rules at the thought level, just as George
Boole believed that “the laws of thought” amounted to formal rules for
manipulating propositions. [Boole, 1854.] I believe that this Boolean dream
is at the root of the slogan “cognition as computation”—and I believe it will
turn out to be revealed for what it is: an elegant chimera.
ENDNOTES TO THE PAPERS
ON ARTIFICIAL INTELLIGENCE
Allen Newell
The four papers in this volume devoted directly to artificial intelligence (AI)
are all quite independent of each other, having been generated by a process
that did not quite live up to the ideal envisioned by the editors, namely, a
main paper plus commentaries. Thus, there is no organic reason for a re¬
sponse. Nevertheless, in the interests of good form, I will make a few com¬
ments. Since structurally they cannot be footnotes, let them be endnotes.
They are just what strikes me as worth noting about the four pieces.
This research was sponsored in part by the Defense Advanced Research Projects Agency
(DOD), ARPA Order No. 3597, monitored by the Air Force Avionics Laboratory under Con¬
tract F33615-78-C-1551. The views and conclusions contained in the paper are those of the
author and should not be interpreted as representing the official policies, either expressed or
implied, of the Defense Advanced Research Projects Agency or the United States government.
287
288 ALLEN NEWELL
analog, not only fits in with the overall picture I drew of the nature of
intellectual issues, but reinforces the existence of several particular issues.
Such quotations are data points for an objective determination of what intel¬
lectual issues existed for AI over time. Finally, Hofstadter’s paper makes
use of an issue of Unitary Objects versus Statistical Aggregates (probably
not quite the right moniker), which coordinates with other issues I did in¬
clude, such as Serial versus Parallel, but is not the same. Hofstadter’s paper
also exemplifies discourse at the level of intellectual issues, but that will
come up in my notes on his paper.
Margaret Boden has provided another essay that continues to establish her
as one of the most knowledgeable and sympathetic critical observers of AI.
The themes she plays out here are ones she has been concerned with for
some time, and their development can be followed in her recently published
collection. [Boden, 1981a.] Especially worthy of attention is her concern
over the issue of whether AI is humanizing or dehumanizing. She argues
clearly for the possibility that AI is humanizing, against the cultural stereo¬
type that equates mechanization with dehumanization. As already noted,
this is an intellectual issue, as that term is used in my paper. Her discussion
reveals an important aspect of behavior with respect to intellectual issues,
namely, reasoning with generalized concepts. Thus, computers are mechan¬
ical, mechanical means nonhuman, therefore computers are dehumanizing,
which makes it difficult for considerations of a different flavor to get a
hearing, such as those that Boden introduces—active intelligence and the
potentialities of computer representation for exploring the nature of subjec¬
tivity. Of course, little progress would be made if one instance of slogan
reasoning simply replaced another, for example, computers have internal
representations, internal representations permit subjectivity, the subjective
is part of the humanistic tradition, therefore computers are humanizing. We
need to find ways to convert discussion at the level of intellectual issues into
scientific studies of the effects (and potential effects) of our artifacts on our
culture.
In this regard, the comments of Boden about the potential relevance of AI
to social psychology need to be underlined. In my own view, the best bet to
unravel the skein of substantive issues about whether complex artifacts,
such as the computer, are humanizing or dehumanizing lies in understanding
in detail social perception and interpretation and the processing that attends
them. Actually, there exists an active, substantial subpart of social psychol¬
ogy called social cognition, whose expressed intent is to bring over into
social psychology modern cognitive theories and results. [Clark and Fiske,
1982; Nisbett and Ross, 1980.] Cognitive psychology is not AI, and this
group is not directly dedicated to bringing over the results of AI into social
psychology. However, AI is one of the sources that created modern cogni-
ENDNOTES TO THE PAPERS ON Al 289
tive theory. I for one would expect the effect of AI on social psychology to
occur largely through such an indirect path.
I seem to be arriving at a view of intellectual issues as having substantial
ideological content—that one of their dysfunctional aspects is that they
permit discourse to stay focused on highly general concepts with ill-defined
semantics. Boden’s comment on another issue reinforces this conclusion,
namely, the psychological relevance of attempts to match experimental re¬
sults against theories expressed in programmed form, which she labels as
debatable. She identifies the basis for the debate to be the uncertain bound¬
ary between what aspects of the program are intendedly psychologically
relevant and what aspects are not. Boden has also made this point in her
early writings about the field. [Boden, 1972.] Indeed, concern with this
problem of the boundary was rampant in the 1960s. [Reitman, 1965, chap. 1.]
My point is that this issue has both ideological and scientific components.
It participates in an intellectual issue about the possible forms for theories of
human behavior—programs versus mathematics versus no-formalism. As an
indicator of its ideological character, the issue of surplus meaning of all
scientific theories has long been noted both negatively (unjustified incorpo¬
ration) and positively (novel implications that prove out). With respect to the
latter, theoretical physicists positively rhapsodize on letting the elegance of
the mathematical formulation lead the way. Yet, the aspect of this perennial
problem that is always brought to bear on AI programs as theories is the
negative side, as if this were a special feature of these AI theories. A little
later in her essay, Boden notes this tendency to apply to AI with a special
intensity concerns that afflict all science. It would seem that this phenome¬
non is not peculiar to AI, but is an aspect of intellectual issues generally.
There are also scientific issues involved here, and genuine advance is
possible in resolving them. In fact, this occurred in the mid-1970s. [Newell,
1973a; Anderson, 1976; Pylyshyn, 1978.] Ideally, an information-processing
theory posits a cognitive architecture for the subject; this is psychologically
significant, just like the posits of any theory. An architecture provides sym¬
bolic structures for specifying behavior—called programs, plans, behavior
schemas, methods, or whatever. Such structures are psychologically
significant even in their details, because the claim that a given architecture is
operative entails the twin claims that (1) the person’s behavior is specified by
such structures and (2) any such structures that can be created by the pro¬
cesses of the architecture can govern behavior. Such a technical advance
provides a principled way of stating exactly the theoretical significance of
the parts of a program proposed as a theory. However, as we saw before, it
will hardly settle the total issue, which is fueled by a larger set of concerns.
I cannot leave Hofstadter’s paper without warning the reader about its
level of scholarship. I will only select one instance, which is of some impor¬
tance both to me and the argument in his paper; numerous additional exam¬
ples, both large and small, occur in the text. Hofstadter devotes a section,
Passive Symbols and Formal Rules, to examining the position of Simon and
myself on the nature of symbols and contrasting this position in stark terms
with his own view (the whole contrast being cast in a peculiarly personal
idiom). Since his paper is available in this same volume, it is not necessary to
quote extensively. However, a quote from the opening paragraph of that
section will help to convey the content of the section:
We can see that to Simon and Newell, a symbol seems to be any token, any
character inside a computer that has an ASCII code (a standard but arbitrarily
assigned sequence of seven bits). To me, by contrast, symbol connotes some¬
thing with representational power. To them (if I am not mistaken), it would be
fine to call a bit (inside a computer) or a neuron-firing a symbol; however, I
cannot feel comfortable with that usage of the term. (Hofstadter in his paper in
this volume.)
The most fundamental concept for a symbol is that which gives symbols their
symbolic character, i.e., which lets them stand for some entity. We call this
concept designation, though we might have used any of several other terms,
e.g., reference, denotation, naming, standing for, aboutness, or even symboli¬
zation or meaning, (p. 156.)
The usual formulations of universal machines also tend to use the term symbol
for the alphabet of distinctive patterns that can occur in the memory medium
(e.g., the 0 and 1 tape symbols for our Turing machine). As defined, these
entities are not symbols in the sense of our symbol system. They satisfy only
part of the requirements for a symbol, namely being the tokens in expressions.
It is of course possible to give them full symbolic character by programming an
accessing mechanism that gets from them to some data structure, (p. 158.)
These quotations are simply the parts of the paper that address directly
and unequivocally the characterization that Hofstadter’s quotation presents.
294 ALLEN NEWELL
The rest of my paper on symbol systems deals at length with the positive
characterization of how symbols represent in general, and how this can
happen by means of finite mechanisms. The treatment is entirely consistent
with Simon’s and my other writings on symbols. The point here is not
whether this theory is known to be right, only that the description of it in
Hofstadter’s paper bears little relation to its actual form.
A FINAL WORD
The only final remark I would make about the collection of papers is that the
historical pictures they all draw of the field seem roughly consonant (though
their evaluations are not necessarily so). To first order, they seem to be
talking about the same history. This is true also of the paper by Pylyshyn
about cognitive science. [Pylyshyn in this volume.] However, I have not yet
seen the other papers in the volume. It will be interesting to see the extent to
which quite different views exist of the total historical framework and to
discover the sources of such differences.
SECTION 4
LINGUISTICS AND ITS
RELATIONS TO
OTHER DISCIPLINES
Thomas G. Bever
Reference
Words refer to objects in the world or mind. Prima facie, such reference is an
obvious and straightforward kind of relation between a symbol and the
world; dog refers to four-footed canines, bachelor to unmarried human
males, and so on. Consideration of the symbol/referent relation has been a
central concern of the philosophy of language, since it bears directly on how
humans represent and know the world around them. The relation between
reference and meaning is not so simple as it seems. We cannot define the
symbolic meanings of words as the sum of their referents, since that would
incorrectly imply that a trivial change in the referent (e.g., in the number of
dogs in the world) would change the meaning of a word. Furthermore,
certain words have no external reference, for example, unicorn, and certain
words seem to be defined only in terms of their effects on other words, for
example, not, did. [Frege, 1879; Katz, 1971.]
Such facts demonstrate that knowing a language involves something other
than simply knowing a large number of symbol/referent associations. This is
also attested by the fact that many animals can master many such associa¬
tions: A skilled shepherd dog can distinguish a large number of different
commands involved in moving a herd of sheep. Yet, we do not conclude that
such a dog has mastered a language.
Propositional Relations
Sentences are the smallest units of language that present well-formed propo¬
sitions that can relate meanings and references. For this reason, current
linguistic investigations focus on the sentence as the major unit of investiga¬
tion. Within a particular language, sentences have distinct structural proper¬
ties. For example, in English, the dog chased the cat means something
different from the cat chased the dog but is very similar in meaning to the cat
was chased by the dog.
The goal of linguistic research is, in large part, to discover the model that
describes sentences and their interrelations. In this essay, I present an out-
LINGUISTICS: RELATIONS TO OTHER DISCIPLINES 299
Rule 1 explains the difference in vowel length between words like hit and
hid, or back and bag. It also explains why the first vowel in bidder is longer
than that of bitter. Rule 2 explains why the second consonant in those two
words sounds alike when they are pronounced rapidly. The ordering of rules
1 and 2 explains why the first vowel in rapidly spoken bidder is longer than
that of bitter, even though the two middle consonants are identical after rule
2 applies to them. The logical ordering of the rules guarantees that the vowel
300 THOMAS G. BEVER
preceding Idl will be lengthened, before the Idl loses its voicing. [Chomsky
and Halle, 1968.]
Levels of Representation
the inner structure of a sentence is the meaning without any further interpre¬
tation. This set up an appealing model in which the inner structure of a
sentence is its meaning, and transformations map it directly onto actual
sequences. [Lakoff and Ross, 1976.] There were a number of difficulties with
this model; most notable was the fact that, as the notion of meaning was
broadened, the corresponding range of possible inner structures had to be
enriched. If two sentences have the same meaning, then on this model it
must be represented by their having an identical inner structure. Simple
examples demonstrate how unwieldy this proposal is. Consider the sen¬
tences “John is an unmarried eligible human male” and “John is a
bachelor.” They clearly have the same meaning, yet construction of a trans¬
formation to relate them involves great complexity.
A further difficulty with the unification of inner structure and meaning
was that certain aspects of meaning appeared to be contributed by surface
structure. For example, different interpretations of one boy ate each jelly
bean and each jelly bean was eaten by one boy are reflected directly in their
surface order. Such examples motivated an expansion of what information
contributes to meaning to include surface structure as well as inner struc¬
ture.
This change occurred at roughly the same time that it was noticed that
outer and inner constituent arrangements are limited to the same con¬
figurations in any given language. [Emonds, 1970.] That is, even though
the possible transformations that apply to inner structures could conceivably
produce many varieties of surface arrangements, most of this power is never
actually used within a language. This motivated the generalization that every
surface structure must be potentially generated by some arrangement of
inner structure constituents. Transformations became correspondingly
simplified to the point at which there is only one transformation, which maps
surface structures from underlying structures, namely, the rule move-X
(where X is either an entire clause or a nounphrase). Conditions on possible
transformations have been enriched, so that the possible surface
configurations are restricted in the right way.
With such developments, it is harder to identify a unique syntactic level
as the primary abstract structure. Viewed intuitively, the first version of the
syntactic model generated surface sequences by applying various transfor¬
mations to the inner syntactic structure; in general, the resulting sequences
were always possible sentences. In its current manifestation, the model
describes many surface sequences that do not correspond to sentences:
They are filtered out by conditions of various kinds, ranging from limitations
on particular surface properties to conditions on possible relations between
surface and inner structures. In this way, the focus of linguistic investigation
has shifted from discovering the transformations to discovering the condi¬
tions of well-formed structures at each level. An important hypothesis (not
yet conclusively proven) is that the conditions are the same for every level of
representation. [Chomsky, 1981.]
LINGUISTICS: RELATIONS TO OTHER DISCIPLINES 303
For several decades, a major concern has been the transformation of infor¬
mation during the first phases of acquiring knowledge. This study involves
the borderlines between processes of attention, perception, and recognition.
When we perceive an object in the world, we use a series of retransforma¬
tions of the stimulus input, starting with the most peripheral sensory repre¬
sentation and moving to the most abstract. We can interpret each stage of
representation as a level of consciousness; the span of attention increases in
their time range and complexity as it shifts from dependence on one kind of
representation to another. Experimental techniques intrude on the ordinary
flow of information, showing which points are under conscious control and
which are not. This provides a picture of consciousness that is somewhat
fragmented. Stable representations of information hop from stage to stage
via brief unconscious periods of transformation.
304 THOMAS G. BEVER
In most cases, the study of this serial process involves information that is
itself of little interest, as the perception and retention of letter groups. Such
stimuli have little internal structure that might yield information on the kind
of transformations that can occur at each stage. A related problem is that the
stimuli do not comprise a natural subset of the world that could call on
psychologically indigenous processes. Language is such a behavior, and
linguistics offers a rich fund of information about possible kinds of represen¬
tations.
The study of how listeners comprehend speech corresponds to the studies
of how information is transformed during immediate encoding. The plausible
view of a retransformation is that listeners first transform the acoustic struc¬
ture into a sequence of separate sounds, which are grouped into words and
then into phrases and sentences and, finally, into meanings.
Careful research on comprehension processes has shown that the succes¬
sive-stage model of encoding is incorrect; rather, the levels of language
representation interact directly during comprehension. We can use our high-
level expectations about meaning to guide our low-level expectations about
sounds. Of course, the points at which such information can be related
between levels depend on the presence of complete representations at the
separate levels. This gives special status to points in a sentence that corre¬
spond to complete units at more than one level of representation; for ex¬
ample, the end of a clause is also typically the end of a meaning proposition
and the end of a constituent. Accordingly, numerous studies show that at
such points listeners integrate the meaning of a sentence with other aspects
of it and its contextual role.
Most important is the fact that sentence comprehension clarifies certain
aspects of consciousness. First, we seem to understand sentences directly as
we hear them, even though our understanding itself modifies the sounds that
we perceive. Various experiments demonstrate that an enormous amount of
computational activity occurs during sentence comprehension, but we are
conscious only of the unified output—the pairing of a sound sequence with a
meaning. This simple fact tells us a great deal about the relation between
separable unconscious processes and their unification within consciousness.
In the case of sentences, the unity of conscious experience can even be
observed in contrast with the duration of the sentence itself. Consider the
sentence you are reading right now. When did it begin as a conscious experi¬
ence as a sentence? After the first word? The second? In fact, it appears to
have begun at the beginning and to have ended exactly at the end. In this
sense, sentences are impenetrable units of conscious experience. Since they
can be quite long, and still be perceived as sentences, one consequence of
language is that it expands the range of our momentary conscious experi¬
ence. Rather than being limited to six items, a sentence expands our momen¬
tary span to 20 words, or 100 sounds! [G. A. Miller, 1956.]
It does so, of course, by utilizing the unconscious mechanisms for rapid
calculation of structural organization. This allows us to use language as a
LINGUISTICS: RELATIONS TO OTHER DISCIPLINES 305
wrong, then the formal theory is wrong. Most scholars took the first course,
assuming that the failure of behavioral predictions based on transformational
analyses was due to the lack of a theory of how the formal theory is utilized
in behavior. This interpretation presumed a distinction between competence
(the formal theory) and performance (behavior). Such a distinction allowed a
retrenchment of the field of psycholinguistics: Linguistic analyses were in¬
voked as describing the structure of the language but not necessarily the
structure intrinsic to language behavior. The role of linguistics was to pro¬
vide structural analyses of sentences. The role of psychology was to explore
which of the analyses actually play a part in speech behavior. [Bever, 1970;
Fodor et al., 1974.]
The distinction between competence and performance seemed clear in the
case of memory, since there are many ways we can choose to memorize and
many different reasons to do so. Speech comprehension and production,
however, offer fewer choices—we generally have one purpose in com¬
prehension, to understand ideas; and generally one purpose in talking, to
express ideas. We might expect that a grammar would express itself more
directly in such behaviors, because there is less opportunity to impose
behavior-specific strategies. However, both behaviors appear to use heurist¬
ics for interrelating sound and meaning: Heuristics are responsive to certain
structural properties of sentences but to other factors as well.
Consider, for example, the usual arrangement of actor, action, and object.
In English, this is the normal order. Various perceptual studies suggest that
listeners impose such an interpretation on sentences as a preliminary (and
unconscious) hypothesis. Clearly, such a strategy is a powerful simplifying
device but has nothing directly to do with a formal grammatical property:
Most clauses do conform to this pattern, but not all.
Despite the clear evidence that certain aspects of talking and listening
depend on nongrammatical patterns and processes, some linguists have
adopted the position that the failure of a grammar directly to predict all
behavioral facts invalidates that grammar. One proposal, in particular, is
that transformations fail to correspond directly to behavioral data because
language is not transformational. According to this view, individual lexical
items carry sufficient structural information to map directly onto meanings.
[Bresnan, 1978.] It is beyond the scope of this paper to present the linguistic
arguments that might bear on such a question. However, it clarifies what
remains an open issue on the relation between formal grammar and speech
behavior: Which properties of behavior directly reflect grammatical knowl¬
edge and which reflect properties of perception and behavior production.
process any skill. In one extreme view, all learning is carried out by a
general-purpose processing mechanism that applies equally to the learning
of language and other abilities. In the alternative view, learning is carried out
by a set of innate mental “modules,” each highly specialized for its task; the
mind would be a federation of such distinct modules.
These two views can agree that the adult mind is highly differentiated into
component capacities; the views oppose each other with regard to the nature
of learning. The computational model was the dominant theory in the psy¬
chology of learning for many years: All learning was held to be the associa¬
tion of stimuli and responses—intelligence consisted of the ability to form
large numbers of such associations. This view was an important rationale for
studying learning in animals, since human mechanisms of skill acquisition
were assumed to be the same, only more capacious. A similar computational
view is now held by many researchers in computer science. The computer is
viewed as a minor brain that can represent the essential features of the
child’s mind. Although the processes may be richer than simple association
of stimuli and responses, the basic claim about learning is the same as that of
the traditional stimulus/response learning theorists: Learning results from
the application of general mechanisms to specific experiences. In particular,
language would be learned as the result of general problem-solving
capacities that children (or computers) apply to the language they experi¬
ence. [Minsky, 1968; Simon, 1969 and 1981.]
Modular theorists argue that speech data the child hears are always too
impoverished to account for what is learned. Therefore, children must have
a particular capacity, innately tailored to learn language; otherwise they
could not extrapolate from limited data to the correct complete knowledge of
their language. The general claim of such theorists is that only artificial and
uninteresting skills are truly learned by the application of general intelli¬
gence. [Chomsky, 1980a.]
This has brought back into psychology an old controversy over whether
the mind is a collection of separate faculties or whether it depends on the
deployment of nonspecific intelligence. All parties to the controversy agree
that some highly specific processes are innate and that there is some capac¬
ity for nonspecific learning. The empirical question involves the nature and
relative importance of the two kinds of processes.
The contrast between general and specific bases for language is central to
biological considerations: Is language the result of a linguistically isolated
evolutionary change, or did it become possible as the result of some general
mutation? Evolutionary speculation about a behavior pattern is limited by
lack of direct records. What we can do is reason from contemporary mor¬
phological and behavioral facts. Three biological properties of language are
308 THOMAS G. BEVER
One aspect of language that differentiates it from other animal activity is the
richness of its acoustic structure. This depends on an articulatory apparatus
that allows for a moving occlusion (produced by the tongue) that separates
two relatively independent chambers in front of and behind it. An ape’s
mouth, for example, simply cannot differentiate the variety of acoustic reso¬
nances that are characteristic of language: The jaw is too recessed to allow
the tongue to separate two large vocal cavities. Paleontological evidence
suggests that relatively recent hominids still had a recessed jaw, which
would not allow for phonological variety. This would mark the evolution of
language capacity as very recent, perhaps less than 100,000 years.
[Lieberman, 1968.]
It is startling to think of the evolution of language as so recent and so
dependent on a peripheral physiological change. Furthermore, the existence
of sign languages in the deaf suggests that vocal expression is not intrinsi¬
cally necessary for language to emerge. One line of investigation concerns
whether sign language is itself a distinct language or whether it is a visual
subset of spoken languages. The research to date suggests that it is an
independent language with rules and properties of its own that are not de¬
pendent on phonetic-based language. Most interesting is the fact that sign
language also appears to respect the more abstract universals of language,
such as the separation of levels of analysis. [Klima and Bellugi, 1979.]
Of course, so far as we know, sign language emerged after spoken lan¬
guage, and may be historically parasitic on it in certain ways. One direct test
of the phonetic basis for linguistic evolution would be to train an animal to
use signs: The most obvious animal is an ape because of its intelligence and
closeness to human beings. If the training were successful, then we would
have some basis for claiming that the particular articulatory developments in
humans might, indeed, have been solely responsible for linguistic capacity.
Teaching apes to talk has, of course, broader implications for our under¬
standing of information and its transmission. A number of recent studies
have attempted to teach chimpanzees sign language or the use of computer-
controlled abstract symbols. [Gardner and Gardner, 1971; Terrace et al.,
1979; Premack, 1976.] The achievements of the chimps are considerable, but
there is no convincing evidence that they learn language structure. Rather,
like shepherd dogs, they can be taught specific sequences and specific rela¬
tional symbols; the capacity to generalize from such training to sentences in
general appears to be lacking. At the moment, such studies have the ambi¬
guity of experimental failures: All we can be absolutely sure of is that exist-
LINGUISTICS: RELATIONS TO OTHER DISCIPLINES 309
ing techniques have failed to teach apes to talk. We can also be reasonably
sure that if some method is found that is more successful, it will involve a
learning environment that is highly contrived. It will be radically unlike the
environment that stimulates language learning in children.
Children acquire the language of the community in which they grow up. This
is obvious evidence in favor of the hypothesis that language is learned. The
acquisition of behavior patterns by animals shows similar learning, for ex¬
ample, a gosling acquires a permanent attachment to its own parent. This
apparent learning turns out to be highly constrained: During a specific phase
of growth, the gosling becomes parentally fixated on any object that has
approximately the right size and the right speed of movement. This is a
typical example of a “critical period,” during which specific experiences
must occur. After the period has passed, the ability of the animal to learn is
closed off. [K. Lorenz, 1965.]
One argument for the biological specificity of language would be that it
also has a critical period. The strongest test of this would be to expose
certain children to language at different ages to see if their learning ability
changes. This ethically unacceptable experiment has been simulated in a few
instances in which children brought up in isolation or by animals are discov¬
ered in adolescence. A recent case study is Genie, a girl isolated until age 11.
Her linguistic ability is still minimal after a number of years of intensive
training. This does not prove that there is a critical period for language, since
many of her other social and cognitive skills are abnormal as well. [Fromkin
et al., 1973.]
Nature occasionally performs a different kind of experiment, in which the
brain is injured during childhood. This gives an opportunity to observe the
speed with which language function recovers. If there were a critical period,
we would expect injury after a certain age to involve categorically greater
loss of language. Clinical surveys have found evidence consistent with this;
injury after age twenty is often followed by long-lasting aphasia, while injury
before age ten is not. [Lenneberg, 1967.] The interpretation of such facts,
however, is not unequivocally in support of a critical period for language¬
learning. First, the critical age would seem to extend for ten years, which is
rather unconvincingly critical. Second, recovery of a function by an injured
brain is not the same as original acquisition of that function. It has long been
known that a seriously injured brain can operate by compensatory functions,
not necessarily replacement of damaged structures with identical ones. [K.
Goldstein, 1948.] Accordingly, the loss of plasticity of neural recovery for
speech may reflect a general loss of plasticity. In particular, it has not been
shown that the loss of language-recovery plasticity at age twenty is any
greater than the loss of plasticity for other capacities of similar complexity.
A supplementary approach has been to explore the evidence for a critical
310 THOMAS G. BEVER
Cerebral Asymmetries
Language is the most salient vehicle for human knowledge and judgment.
Accordingly, it is often treated as central to issues involving the philosophy
of mind. Indeed, certain periods in philosophy have been dominated by the
view that in solving linguistic problems, we can solve fundamental problems
concerning epistemology and truth. Modern linguistics bears on three main
issues: the biological basis of knowledge, reductionism as the standard form
of scientific explanation, and the existence of Platonic a priori structures.
There are numerous properties P that might be invoked. The fact that differ¬
ent kinds of knowledge contribute to the usability of a sentence itself might
be undiscoverable from actual data given. How does the child actually learn
that certain uses of a sentence are acceptable for pragmatic as opposed to
312 THOMAS G. BEVER
structural reasons? It seems unlikely that all children receive explicit train¬
ing in the different ways that sentences can be used. No explicit information
tells a child that sentences can be used ironically to have the opposite of
their ostensible meaning, yet the child acquires this distinction.
Linguists tend to cite more technical properties of linguistic structure as
unlearnable by standard learning theories. For example, the restriction that
transformational rules must apply to the highest level constituent of a given
type is never explicitly presented to the child. Clearly, the child must apply
some pattern-recognition strategies to the data as given and impose on them
the presumption that such a condition exists. But this imposition itself can
draw on only hypotheses internal to the child: Hence, the concept of such a
condition on rules must be an innate hypothesis.
The accumulation of such arguments adds up to the claim that the inter¬
esting formal properties of language are all innate. To paraphrase Chomsky,
there is a “language-learning organ” that can extrapolate a full grammar
from extremely scanty data. Language is not learned but rather grows in the
mind, much as the liver or kidney grows in the body. [Chomsky, 1981.]
This nativist claim, of course, is itself open to empirical investigation:
What children have learned when they use language appropriately is an
object of psychological inquiry. Chomsky, however, puts the nativist claim
in a context that confronts empiricism in its broader sense: He extrapolates
from the relatively narrow claims about language development to broader
claims about all human knowledge. He suggests that there is a general
theory-forming capacity that is itself an innate mental organ.
The clarity of this attack on empiricism has created a rare situation. A
philosophical issue involving the nature of knowledge might actually be
resolved in part by appeal to empirical data. In this case, critical data are the
patterns of language acquisition: If children can be shown to use gradually
some general pattern-learning skills to learn language, then the specific
nativist claim is weakened; conversely, insofar as children can be shown to
extrapolate creatively beyond speech data in ways that express idiosyncratic
formal linguistic universals, the nativist claim is strengthened. At the mo¬
ment, the weight of evidence from language acquisition is in favor of the
nativist position. This, however, may be related to the fact that no strong
theory of pattern acquisition has been proposed that might explain language
acquisition as a special case. No one has yet shown that such a theory is
impossible in principle. [Slobin, 1971 and 1979.]
APPLICATIONS OF LINGUISTICS
and theoretical problems. Today, computers are the most salient devices for
practical implementation of linguistic models. In the classroom, such prob¬
lems as language-learning and reading involve a mixture of theory and prac¬
tice. Finally, linguistics has been taken as a model for studying other human
symbolic activities.
Computer Science
One of the frequently proposed tests for the adequacy of a computer model
of human thought is whether we could tell that it were not human simply by
interacting with it through language. If the computer model responds in a
humanly plausible manner, success has been achieved. [Turing, 1936.] This
operational test is surprisingly easy to achieve if the communication context
is restricted. For example, a well-known computer program simulates the
conversational activity of an analyst. This program is effective just because
the analyst it simulates provides little new information in a therapy session.
Rather, the program reformulates statements typed in by the user and turns
them into corresponding questions. This leads us to conclude that the pro¬
gram is not a successful replication of a human being, even though it might
be hard to detect the difference in the restricted context.
We must treat every language capacity in a computer relative to its appli¬
cation. Three major kinds of applications have been attempted during the
past few decades: automatic translation, automatic comprehension, and
special-purpose implementations.
Automatic translation would seem to be an obvious candidate for com¬
puter processes. After all, much of translation involves word-by-word map¬
ping from one language to another. Computers can certainly implement
bilingual dictionaries with great speed. This does not, however, guarantee
good translation; in fact, it often produces completely incomprehensible
sequences in the target language. This is partly because words in sentences
represent meaning by virtue of their interaction together; partly because
there are numerous multiword idioms in every language; and because cor¬
rect interpretation of a sentence often involves an understanding of its rela¬
tion to other sentences in the same discourse and to the nonlinguistic context
in which it is used. Human translators bring their knowledge of the world to
bear as they translate, something that a computer cannot do, no matter how
sophisticated its linguistic program.
Machine translation dominated attempts to apply computers to language
for several decades, but there has been no general solution. Limited suc¬
cesses can transfer sentences in scientific documents from one language to
another, but human readers are expected to compensate for the occasional
mistakes that appear in even this highly restricted context. There is no
reason, however, why moderately good translation should be impossible in
principle. As computers become more powerful and have more storage,
“brute force" solutions become possible, that is, solutions that involve a
LINGUISTICS: RELATIONS TO OTHER DISCIPLINES 315
achievable. Of course, here, too, the perfect solution would involve under¬
standing the utterance; but much of practical value can be achieved by
computer-generated speech, just because human listeners compensate for its
imperfections.
George A. Miller
*1 am indebted to Dan Sperber for suggesting this way of stating the alternative view.
319
320 GEORGE A. MILLER
information required to keep this system functioning is not yet well under¬
stood, but some theorists have spoken, not entirely metaphorically, of the
languages of the brain. The basic communicative process is self-regulatory,
not social.
According to this view, the unique event in human evolution occurred
when our ancestors somehow became able to use the languages of the brain
for social communication. Some mutation made it possible to externalize the
internal information-processing—some kinds of internal processing, at
least—in vocal noises. Thus, human patterns of thought did not await the
evolution of language but preceded it. Social cooperation based on the frag¬
mentary clues to thought normally expressed in human speech would sel¬
dom succeed if the internal information-processing were not closely similar
from one individual to the next; such uniformity is not characteristic of the
environments in which people live, but it could be achieved genetically. In
short, language is what it is, not because social interactions are what they
are, but because the brain is what it is.
People frequently have trouble understanding spoken messages that they
do not expect, even when the acoustic signal is loud and clear. The reliance
on expectations is sometimes called top-down processing of speech, by
contrast to bottom-up processing that begins with the acoustic signal, seg¬
ments it, then filters and classifies the segments. Engineers who have tried to
build speech-recognition systems have had their greatest success with bot¬
tom-up approaches, from which it is sometimes hastily concluded that the
importance of top-down processing has been vastly overrated. A more pru¬
dent conclusion would be that engineers have not yet been able to simulate
top-down processing very well, a conclusion that would follow from decent
respect for the complexity of neural information-processing and our ignor¬
ance of how it is done. Everything a person knows about another person’s
situation, past or present, can be used to limit what that other person could
be expected to say and so make perception of what they do say easier. That
kind of coordination is possible because different people think the same
way, but no machine has yet been programmed to think as people do.
Part of the internal information-processing that all people do by virtue of
being human is described by the universal properties and principles of levels
of syntactic representation. The good news is that it has been possible to
characterize even a fragment of this internal processing. The bad news is
that that is but a small part of the system that psychologists eventually hope
to understand.
CURRENT VIEWS OF
LANGUAGE AND GRAMMAR
Ray C. Dougherty
FOUR POSITIONS
One can obtain an understanding of the basic problems, if not the solutions,
by considering recent works in which four distinct positions have been dis¬
cussed. One position, the empiricist view, characterized by one author cited
later as a mere recording of observations without a structuring activity on
321
322 RAY C. DOUGHERTY
the part of the subject, governs little current research. Each of the three
remaining positions has its champions.
Chomsky, who has formulated most of the issues defining the biological
perspective, presents an “innatist,” or, as Piaget calls it, “preformationist”
view, according to which all of the basic linguistic structures are precoded
into the genetic constitution of a human being and triggered by experience.
Jean Piaget presents a “constructionist” view, which claims that the
basic properties of a grammar are neither implicit in the data to which a child
is exposed nor innate in the mental organization of a child. Constructivism
assumes that grammatical structures emerge as constructions developed by
general sensorimotor processes when applied to language data. A child
learning a language essentially creates a grammar anew by general processes
of sensory perception and motor coordination that are not language-specific.
Seymour Papert has recently developed a computational approach, called
Project LOGO, which is closer to Piaget’s constructivism than to
Chomsky’s innatism. Papert has attempted to develop a computer-based
model of learning that uses mechanisms developed in computer science to
characterize neurological structures. Papert developed the computer lan¬
guage LOGO in order to model the information processes that function to
characterize knowledge and the mechanisms by which knowledge is repro¬
duced in a new brain. Papert assumes that one could better understand the
processes of grammar and language if one had more understanding of the
nature of machines that learn and machines that learn how to learn. He uses
LOGO to design such machines and model the neurological processes in¬
volved.
mental properties of the eye or the visual cortex or the heart develop on this
basis. Furthermore, when we turn to specific properties of this mental organ,
we find little justification for any such belief, so far as I can see. (Chomsky,
1980fl, p. 37.)
Chomsky believes that the structures of language and the complex pro¬
cesses of grammar development could not arise from organism/environment
interaction alone. Grammar is species-specific and genetically determined.
Chomsky elaborates:
I have touched on only a few examples. In each case, when we investigate the
particular properties of human cognition, we find principles that are highly
specific and narrowly articulated, structures of a marvelous intricacy and deli¬
cacy. As in the case of physical organs, there seems to be no possibility of
accounting for the character and origin of basic mental structures in terms of
organism-environment interaction. Mental and physical organs alike are deter¬
mined, it seems, by species-specific, genetically determined properties, though
in both cases interaction with the environment is required to trigger growth and
will influence and shape the structures that develop. Our ignorance—
temporary, let us hope—of the physical basis for mental structures compels us
to keep to abstract characterization, in this case, but there is no reason to
suppose that the physical structures involved are fundamentally different in
character and development from other physical organs that are better under¬
stood, though a long tradition has tacitly assumed otherwise. (Ibid., pp. 51-
52.)
At the same conference where Chomsky offered these views, Piaget defined
constructivism as an alternative to both empiricism and innatism (or prefor-
mationism). Said Piaget:
Fifty years of experience have taught us that knowledge does not result from a
mere recording of observations without a structuring activity on the part of the
subject. Nor do any a priori or innate cognitive structures exist in man; the
functioning of intelligence alone is hereditary and creates structures only
through an organization of successive actions performed on objects. Conse¬
quently, an epistemology conforming to the data of psychogenesis could be
neither empiricist nor preformationist, but could consist only of a constructiv¬
ism, with a continual elaboration of new operations and structures. The central
problem, then, is to understand how such operations come about, and why,
even though they result from nonpredetermined constructions, they eventually
become logically necessary. (Piaget, 1980, p. 23.)
Piaget defines what he sees as the basic problem of research into the
nature of language and grammar: To choose between constructivism and
324 RAY C. DOUGHERTY
These few examples may clarify why I consider the main problem of genetic
epistemology to be the explanation of the construction of novelties in the
development of knowledge. From the empiricist point of view, a “discovery”
is new for the person who makes it, but what is discovered was already in
existence in external reality and there is therefore no construction of new
realities. The nativist or apriorist maintains that the forms of knowledge are
predetermined inside the subject and thus again, strictly speaking, there can be
no novelty. By contrast, for the genetic epistemologist, knowledge results from
continuous construction, since in each act of understanding, some degree of
invention is involved; in development, the passage from one stage to the next is
always characterized by the formation of new structures which did not exist
before, either in the external world or in the subject’s mind. The central prob¬
lem of genetic epistemology concerns the mechanism of this construction of
novelties which creates the need for the explanatory factors which we call
reflexive abstraction and self-regulation. However, these factors have fur¬
nished only global explanations. A great deal of work remains to be done in
order to clarify this fundamental process of intellectual creation, which is
found at all the levels of cognition, from those of earliest childhood to those
culminating in the most remarkable of scientific inventions. (Piaget, 1970, pp.
77-78.)
in discussing one of his examples, indicates how Piaget influenced his work:
“When I read Piaget this incident served me as a model for his notion of
assimilation, except I was immediately struck by the fact that his discussion
does not do full justice to his own idea.” (Papert, 1980, p. vii.) Papert
mentions that Piaget’s ideas gave him “a new framework” for examining his
observations. (Ibid., p. viii.) Papert’s project LOGO is a version of construc¬
tivism and opposed to both empiricism and preformationism. Papert states:
It seems to be beyond doubt that the brain has numerous inborn “gadgets.”
But surely these “gadgets” are much more primitive than is suggested by
names like Language Acquisition Device and Math Acquisition Device. I see
learning language or learning mathematics as harnessing to this purpose numer¬
ous “gadgets” whose original purpose bears no resemblance to the complex
intellectual functions they come to serve. (Papert, 1980, p. 220.)
EXPERIMENTAL ANALYSIS
Much of the excitement of research into the structure of language and the
genetic processes governing grammar reproduction derives from the fact
that the results of research bear directly on traditional questions about the
structure of knowledge and the mechanisms by which it is communicated
from the older to the younger generations. In linguistics, traditional philo¬
sophic questions can be posed in sufficient detail that they can be subjected
to experimental analysis.
RULES AND PRINCIPLES
IN PHONOLOGY AND SYNTAX
The following piece attempts to do two things. First, it attempts to show why
linguists suppose that mastery of a language is tantamount to mastery of a
system of rules rather than, say, memorization of a very large number of
words, phrases, sentences, and so on. Second, it draws attention to a major
difference that has emerged in recent years between the rule systems of
phonology and the rule systems of syntax and speculates on the significance
of this difference.1
WHY RULES?
A fundamental tenet of linguistic theory is that what people know when they
know how to speak is a system of rules. The argument goes as follows:
1 The author is indebted to Morris Halle and James Higginbotham for a discussion of some of the
issues raised in this piece.
327
328 SAMUEL JAY KEYSER
3. The argument in (1) shows that there can be no upper limit on the
number of sentences in English. However, we know that the human
brain, the repository of linguistic ability, is finite. Even if we were to
assign one sentence to each atom of gray matter in the brain, we
would run out of atoms long before we ran out of sentences.
4. If knowledge is represented mentally and if we have the ability to
produce an infinite number of sentences, then how can we represent
this infinite knowledge in a finite space?
5. The answer is to assume that the knowledge that enables us to pro¬
duce an infinite number of sentences is not a representation of each
individual sentence, but rather a set of rules, finite in number, whose
output is, in principle, infinite.
6. Therefore, what we know when we know how to speak is, among
other things, a set of rules.2
We have seen in syntax that rules are required since speakers know, in
principle, an infinite number of sentences. However, in phonology we might
argue that since the number of morphemes and words is finite, there is
nothing incoherent about supposing that a speaker simply memorizes the
phonetic makeup of each word. How might this work?
Consider the well-known fact that regular English plurals are pronounced
in three separate ways: either as the syllable [iz] as in kisses or as the
consonant [s] as in hits or as the consonant [z] as in dogs. We might say that
when a child learns the English plural, the child simply memorizes these
facts. In other words, just as a child must remember that the plural of mouse
is mice, the child must also remember that the plural of spouse is spouses
and not, say, spice. Such a model is theoretically possible, since the number
of words in English is finite.
2This argument makes use of that property of natural language that Bever calls “the creative
component of sentence structure.”
RULES AND PRINCIPLES IN PHONOLOGY AND SYNTAX 329
One way to argue for the view that phonology is rule-governed is to show
that novel forms do not present a problem for speakers. That is to say,
speakers are able to treat novel forms as if they were not novel at all.
Suppose, then, we were to ask someone to provide us with the plural form of
a word that ended in a consonant sound that was not a normal English
consonantal sound. Thus, consider the sound [xL which terminates the
name Bach', that is [bax].3 Even though this sound is not an English sound, a
native speaker of English would always pronounce the plural of this word
with an [s] and not with an [iz] or a [z]. This response can easily be verified
by asking native speakers of English whether they have ever heard the word
before and, if not, what they suppose its plural to be. Moreover, this re¬
sponse can only be understood if we assume that part of a native speaker’s
knowledge of the English plural is knowledge of a rule and not knowledge of
pairs of singular and plural forms. Thus, if there were no rule, it would be an
incredible coincidence that all English speakers treat the plural of Bach as [s]
and not, for example, [baxiz]. Notice that both forms are easily pronounce¬
able in English, and, were there no rule, we would have to assume that at
least one speaker would opt for [baxiz]. This, in fact, never happens.4
We saw that in syntax the argument in favor of rules was based on the
need for an infinite output. We have just seen how in phonology the argu¬
ment in favor of rules is based on novel forms. However, both of these
arguments are, in fact, based on the same property, namely, novelty. Thus,
for speakers who have never heard the name Bach before, the name consti¬
tutes a novel word whose plural will, nonetheless, be obvious to them.
Similarly, a sentence patterned after those in the preceding example that is,
say, three times as long as (d), or even ten times as long, will be a novel
sentence to anyone with the patience to put such a sentence together. None¬
theless, such a sentence, despite its novelty, will easily be recognized as
being a perfectly grammatical one in English by anyone who knows the
language.
It is not hard to show, however, that the premise on which the ruleless
hypothesis is based, namely, that the number of words in English is finite, is
itself incorrect. Consider, for example, the words lion and hunter. English
3This is the voiceless, velar sound that occurs in German words like ach.
4We will not pause here to provide a formal statement of the English plural rule. However,
several accounts can be found in contemporary literature. For one that also includes an account
of its historical evolution, the reader is referred to Keyser and O’Neil [1980].
330 SAMUEL JAY KEYSER
has a process called compound formation whereby these words can be put
together to form a single new word, lion hunter meaning a hunter who hunts
lions.5 Notice that if we can add a word to the word hunter once, there is no
reason why we cannot perform the same operation twice. Thus, suppose we
now add the new word lion hunter to the word hunter to produce another
new word, namely, lion hunter hunter. This is not only an acceptable word
in English, it is also perfectly clear what it must mean. A lion hunter hunter
means a hunter who hunts lion hunters.
We have now constructed an argument that is formally identical with that
in the first set of sentences. We must have phonological rules because our
knowledge of words, like our knowledge of sentences, is infinite. This is the
novel-form argument for rule-based linguistic systems. Indeed, all novel
forms can be construed as evidence in favor of the view that rule systems
underlie our linguistic competence.6
Even though syntax and phonology share the property of creativity (see
footnote 2), there is an interesting difference between the two components.
In his subsection on “General and Specific Learning Mechanisms,” Bever
makes reference to modular versus general purpose models of learning. The
modularity hypothesis assumes that “learning is carried out by a set of
innate mental ‘modules,’ each highly specialized for its task . . . .” In this
view, the mind is a “federation of such distinct modules.”
The modular model of mental representation describes the current state of
affairs in linguistic theory as well. This change toward modularity in lin¬
guistic theory has come about over the past few years. Noam Chomsky, in
his recent monograph, puts the matter in the following terms: “There has
been a gradual shift of focus from the study of rule systems, which have
increasingly been regarded as impoverished (as we would hope to be the
case), to the study of systems of principles, which appear to occupy a much
more central position in determining the character and variety of possible
human languages.” (Chomsky, 1982, pp. 7-8.)
5 Even though we tend to write such compound words with space between them, this ortho¬
graphic convention should not mislead us into thinking that we are not dealing with a single
word. For one thing, this word takes plural endings just like other single words. Just like other
single words, we cannot put things inside it. Thus, words like taffeta cannot be interrupted and
neither can words like lion hunter. Hence, we cannot say lion big hunter even though we can
say big hunter. Moreover, we cannot place the plural ending inside the word; that is, the plural
of lion hunter is lion hunters and not lions hunter. We see, then, that in a number of ways lion
hunter behaves like a single word.
Novel forms come from a variety of sources. Errors that speakers make is one common
source. Thus, when a child assumes that the plural of mouse is mouses and not mice, the child’s
behavior can best be understood as resulting from an overgeneralization of the rule for the
regular formation of plurals in English.
RULES AND PRINCIPLES IN PHONOLOGY AND SYNTAX 331
Note, first, that (b) is related to (a). In particular, (b) contains a fronted
question word what. However, while this word appears at the front of the
sentence, it is, nonetheless, the object of the verb buy. To capture this
intuition, we place an abstract element called a trace (= t) where the object
of buy normally appears, and we index the question word and the trace to
indicate the relationship that exists between them. This is indicated in (c),
which therefore constitutes the appropriate representation of the question
corresponding to (a) at one level of grammatical representation.8 Consider,
now, the question that corresponds to (d) at this same level of representa¬
tion, namely, the sentence indicated in (e). Here, too, a trace element relates
the fronted question word what with the position that the object of buy
normally occupies. This brings us to the final pair of sentences, namely (f)
and (g). Notice that whereas (f) is a perfectly well-formed sentence in En¬
glish, it does not have a corresponding question form of the pattern indicated
in the earlier questions, namely, in (c) and (e). When we attempt to repro¬
duce that pattern, we get an ungrammatical sentence such as that in (g). On
the surface of things this is a remarkable fact. Why, after all, should it not be
possible to question (f) just as it was possible to question (a) and (d)? To
account for this fact, a principle is assumed to apply at the level of structure
corresponding to (g). This principle, called subjacency, states that a fronted
element may not occur too far away from its original position (where, for
purposes of the present exposition, we leave unspecified what too far away
means formally):
7The shift from rules to principles in no way undermines the argument in (1) that knowledge of a
language means knowledge of a grammar. Rather the argument goes through but with principles
replacing rules as the dominant mechanism of grammatical knowledge.
8This level is called 5-structure, and it corresponds roughly to the notion of surface structure.
332 SAMUEL JAY KEYSER
This brief enumeration barely does justice to the active character of the phonological com¬
ponent.
The velum is the soft fleshy valve at the roof of the mouth, which, when shut, closes off the
nasal cavity and, when open, allows the nasal cavity to resonate.
RULES AND PRINCIPLES IN PHONOLOGY AND SYNTAX 333
CONCLUSION
I conclude with a question. Higher levels of grammatical theory have not yet
achieved the marriage of the formalism with the mechanism that phonology
has at least begun to achieve. As we learn more about mental mechanisms,
will syntactic and semantic theory change in ways that will reflect those
mechanisms just as phonological theory has done, or is the present nature of
syntax and semantics reflective of those mechanisms already?11 If a final
speculation is in order with respect to this piece, then that speculation would
be that syntax and semantics will follow phonology as more is learned about
higher order mechanisms of the brain. Given the rightness of this specula¬
tion, we can expect radical changes in syntactic and semantic theory to
confront us in the future.
POSTSCRIPT
Before his sudden and unexpected death, Fritz Machlup requested that I
might add a word about the relationship between my remarks here and
studies in the information sciences such as those represented in the present
volume. At the risk of going beyond the space limitations imposed on me by
the editors, I would like to attempt to comply with his request.
During the past quarter of a century, the study of linguistic theory has, to
a great extent, been carried on independently of the study of the computa¬
tional mechanisms that enable the sentences of a language to be both pro¬
duced and recognized. Thus, linguistic theory studies human language com¬
petence, while the study of computational mechanisms, which might be
"James Higginbotham (personal communication) has suggested that the static character of
higher level components has resulted from a shift away from conditions on rules (a dynamic
device) to filters, or conditions on well formedness at a linguistic level (a static device) as a
mechanism of grammatical description. This shift has come about for reasons internal to lin¬
guistic theory itself; in particular, because of the ease of description that results from the use of
filters as opposed to rules. Phenomena under current study would require significantly more
complicated statements if formulated in terms of rules rather than filters.
334 SAMUEL JAY KEYSER
A view widely held by academics over the past few decades has been that
computer science and modern linguistic theory do not mix. Several years
ago, so the story goes, linguists driven by Noam Chomsky’s distinction
between linguistic competence and linguistic performance proposed that the
study of what people know about their language should take precedence
over the study of how people actually process or produce utterances or
actually learn language. This distinction between what and how has attained
a kind of self-perpetuating status as the driving wedge between linguistics on
the one hand and artificial intelligence and computer science on the other—
see Newell’s characterization of it as a “field-splitter.” [Newell in his lead
paper in Section 3 of this volume.]
But times change. I would like to argue that we would be hard pressed to
find a more complementary marriage of disciplines than that between lin¬
guistics and computer science and that such a marriage has recently borne
and will continue to bear great fruit. Linguistics and computer science are
partners of a very special kind, because the study of human cognition is now
widely equated with the study of information-processing models. But, as
every computer scientist knows, such models consist of algorithms that
juggle very specific information structures. Algorithmic procedures for using
knowledge do not make sense without a specification of the form in which
that knowledge is couched. Similarly, an understanding of how language is
used hinges on an understanding of what that knowledge looks like—the
way it is represented. So the joint work of linguistics and computer science
is like the partnership between data structures and the algorithms that use
them. In the remainder of this paper I would like to make plain just what the
nature of that partnership is and what one can expect to gain from it.
When talking about a living organism—any organism—there are two
broad kinds of “how” questions, and each has its place in the explanation of
human information-processing abilities. The first of these questions, raised
in Dougherty’s paper, is couched at the level of the species: How is what the
individual organism knows (a body of information) passed on to its descen-
335
336 ROBERT C. BERWICK
dants? Or, to put the question from the perspective of the offspring, how is
knowledge of language acquired? This is a particular kind of performance
that is rarely mentioned by researchers concerned with the analysis of the
day-to-day use of language—for example, it is not even mentioned in
Newell’s lead paper in this volume as a goal of the computational study of
linguistic performance. The second kind of how is, in fact, just this more
familiar one of individual language use: How do people actually use informa¬
tion about language to speak or understand sentences?
The linguist’s answer to both questions, as exemplified in the papers by
Bever, Dougherty, and Keyser, is to characterize what an individual knows
about language as a system of rules, a grammar. As an example of what a
part of a grammar looks like, consider the following sentences (like those
described in Keyser’s paper):
Now, as Keyser observes, what modern linguistic theory does is tell one that
(b) really has the underlying form,
where [t] acts like a variable bound to the value of who, so that the question
is really to be construed as,
Further, what people seem to know about forming questions is that one can
take a statement like (a) and turn it into a question by imagining that who is
the object of like:
and then moving who to the front of the sentence, inverting the auxiliary
verb and subject, and leaving behind the (unpronounced) variable X. Appar¬
ently, though, we cannot move who very far. We can move it out of a lower
sentence (for example, Mary likes who) into the next higher sentence but no
farther. Look what happens when we try to move who out of a sentence that
is in turn embedded in a nounphrase, the claim that . . . :
The sentence is ill-formed, and we know this, even though we can make
out what it would mean if we could say such a thing. This is part of our
COMPUTATION DOES MATTER TO LINGUISTICS 337
knowledge of English, part of our grammar. Because this knowledge has the
effect of restricting the radius of action of the movement of who, we might
call it a locality principle, PI:
of all three levels has been determined, then we can use what we know about
the implementation and algorithmic constraints to tell us something about
the abstract problem, and vice versa.
Turning now to the situation of language, the first thing we notice is that
what plays the role of the abstract computational problem are forms such as
(c). If one thing that we do in order to understand a sentence is to recover its
underlying form, then we must convert the string of words,
That is, linguistic theory tell us what the abstract computational problem is,
since it tells us what the units of representation are that must be recovered.
Now let us consider the how of it all. Just as with the 100-integers problem,
there could be many ways of actually performing the computations required
to get to the underlying form. And there is one additional factor to consider:
While we are free to speculate about algorithms and implementation
methods to solve the integers problem, in the case of linguistic algorithms
and implementations we are shackled not by our own creativity but by
Nature’s—at least if we want to explain what it is that people do. The chains
are wound doubly tight because what we know about the machinery that the
brain has at its disposal is quite limited; we know that it does not have add
and multiply buttons, but we know very little more than that.
Even so, we can still advance computational arguments for and against
certain ways of characterizing the what of linguistic theory. Let me give just
one example of this style of argumentation here, an example showing that
there is some computational evidence for the abstract form (c) posited by
modern transformational grammar. In passing, the argument will illustrate
one way in which a theory of computation can support the study of linguis¬
tics.
To begin, consider the following two algorithms (machines) for analyzing
sentence (d). Each machine takes as input the sentence to be analyzed and
writes as output a desired underlying form corresponding to the input sen¬
tence.
Algorithm 1. Imagine sliding a window across the input sentence. De¬
pending on what it sees in the window, the analysis machine makes a deci¬
sion as to what output to produce next. The window is fixed in width, in the
sense that it can hold a finite number of words (such as did) or phrases (such
as a sentence or nounphrase) but no more. This means that we can express
the operations of the machine in terms of finite patterns and actions, where a
pattern is simply a finite string of words or phrases, and an action is what the
machine writes as output. In our example, if the window of the machine
holds the items
COMPUTATION DOES MATTER TO LINGUISTICS 339
then the machine writes out the same string as it sees, with the [t] marker
inserted as desired
W. Boyd Rayward
The purpose of this paper is to explore some of the relations between librari-
anship, bibliography, documentation, library science, and information sci¬
ence. All are considered part of a historical process that has led to different
ways of envisaging, creating, and investigating the interrelations and relative
effectiveness of formalized modes of access to recorded knowledge.
Underlying this approach are two observations. First, libraries represent
long-established and complex arrangements of a variety of these formalized
modes of access to knowledge. Second, not long after librarianship emerged
as a profession whose practitioners were chiefly absorbed by the adminis¬
trative and operational challenges presented by libraries, there arose a
countervailing resistance to the conceptual narrowness that such a profes¬
sional preoccupation involved. Important as libraries had become, histori¬
cally they represented only a partial solution to the constellation of problems
related to the conservation, organization, and diffusion of recorded knowl¬
edge and information. If libraries were to be improved or if other solutions to
the problems for which they presented partial solutions were to be invented
and tested, then it seemed necessary continually to seek new ways of look¬
ing at libraries themselves and at the broader environment of which they are
only a part. Librarianship, bibliography, documentation, library science,
and information science may be considered as incorporating modes of study
and investigation that not only express subtle occupational distinctions, but
also represent attempts at obtaining these new and increasingly general per¬
spectives.
In the second part of this paper, I examine five developments in the field
that have occurred in the United States in the last ten to fifteen years. I
suggest that these developments, relatively insignificant in themselves, are
important as indicators that librarianship and information science, the latter
arising in part from the documentation movement, have certain formal con-
343
344 W. BOYD RAYWARD
nections. These connections may well imply that there has been a disciplin¬
ary convergence of librarianship toward information science. The develop¬
ments to be examined in this connection are some recent changes in the
names of library schools; the employment of librarians in information cen¬
ters and other less conventional contexts; the growth in the number and the
diversification of the specialized content of journals in librarianship and
information science; the appearance of an important body of new terminol¬
ogy in librarianship; and finally, the formal structural recognition of mutual
interests that has taken place within the major professional associations in
librarianship and information science.
In relating the first part of the paper to the second, I take the somewhat
simplistic view that the prominence of the computer in the last ten or fifteen
years has given librarians themselves a new power of seeing beyond the
conventional boundaries of their thinking and activities traditionally im¬
posed by their concern for individual libraries. Because of the increasing
sophistication and accessibility of computer technology, librarians have
been able to develop library and information services and to organize, man¬
age, and exploit library and information resources in a manner and to an
extent not possible before.
It is also my view that information science—almost a phenomenon of the
computer age—represented in part, at least initially, systematic attempts to
generalize from and to explore in as rigorous a manner as possible questions
and answers that were expressed in what had become accepted library or¬
ganization and practice. This was not at first clearly recognized or at once
accepted.
In the third and final part of this paper, asking a question posed by Don
Swanson [1980/)]—What are the problems libraries were created to solve?—
I try briefly to show that there is a disciplinary continuum between librari¬
anship and information science with no easily identifiable boundary separat¬
ing them, though the difference between the extreme ends of the continuum
are clear and even dramatic. I also suggest that the disciplinary movement
toward more general and inclusive perspectives discussed in the historical
part of this paper may still continue to express itself in a tendency to seek a
redefinition of the objects of study of information science.
Underlying this paper is a simplification, which emerges in various ways,
that should be acknowledged at the outset. I consider the formation of
stable, long-lived scholarly and professional associations and societies as
key events. They are culminations in a developmental process of differentia¬
tion in librarianship, bibliography, documentation, library science, and in¬
formation science that I single out from a complex history that involves
much more than them. On the whole, I view the fields designated by these
appellations in terms of the occupational characteristics, and a related schol¬
arship, of various organized groups of individuals. I focus on some of the
changing structural and organizational expressions of relations between
these groups and the scholarship to which they lay claim, rather than on the
LIBRARY AND INFORMATION SCIENCES 345
LIBRARIANSHIP
BIBLIOGRAPHY
But within a generation, it became clear that neither libraries nor the existing
organizations of librarians were able to initiate or even contribute usefully to
the study of solutions to certain general bibliographic problems that lay
346 W. BOYD RAYWARD
outside the walls of these individual libraries. This was as true in the United
Kingdom as in the United States. Thus, the creation of the Bibliographical
Society in 1892 was proposed and discussed at the Library Association
meeting in 1891. An issue debated at length was just what the relation of the
Bibliographical Society of Chicago, founded in 1899 (to become the Ameri¬
can Bibliographical Society in 1904) would be with the American Library
Association (ALA); indeed, the fourth general session of the 1902 confer¬
ence of the ALA was placed “in charge” of the Bibliographical Society of
Chicago for a discussion of the founding of a Bibliographical Institute.1
These several bibliographical societies were interested in bibliography as
an intrinsically rewarding area of study that had produced notable achieve¬
ments in its already distinguished history. Bibliography as these societies
conceived it had broad consequences for scholarship quite apart from librar¬
ies. In England, there was in the early days of the Bibliographical Society
the intoxicating idea of such a society creating a great catalog of English
literature and preparing a revision and supplement to Hain’s Repertorium
Bibliographicum. [Copinger, 1892; Ray ward, 1967a.] In both England and
the United States, membership in these early bibliographical societies com¬
prised, certainly at first, many librarians but also, and perhaps predomi¬
nantly as time went by, book collectors, antiquarians, professional literary
and historical scholars, and amateur men of letters who incorporated the
results of their specialized research in arcane papers that were com¬
municated to each other at meetings and in the journals of their societies.
Gradually, interest in what we would today call general issues of biblio¬
graphic control tended to subside in these societies, though never com¬
pletely in the American Bibliographical Society, to be replaced by a schol¬
arly preoccupation with books as physical objects.
The history of books, printing, and publishing, the description of old and
rare books, special collections of books and manuscripts, the implications of
the physical characteristics of codices and books for various disciplines, and
perhaps above all problems in the transmission of texts, such as the relation
of manuscript to printed versions—these gradually became the preoccupa¬
tion of the members of these societies. As specializations of a highly techni¬
cal kind developed, bibliography unadorned became too vague, too much a
portmanteau term to be useful. Nowadays, the scholarly activity of bibliog¬
raphers is labeled descriptive, historical, analytical, or critical bibliography
according to particular, perhaps not precisely or consistently distinguished,
criteria of interest, affiliation, and technique.
Bibliographers asserted the overall independence of their own area of
intellectual endeavor in new journals, among which The Library in Great
Britain, the Papers of the Bibliographical Society of America in the United
States, or, later, the Proceedings of the Bibliographical Society of Virginia
are important. Such journals help to define the boundaries between biblio-
DOCUMENTATION IN EUROPE
LIBRARY SCIENCE
In the United States, the situation was a little more complex. Here, dissatis¬
faction with traditional practices of librarianship led not merely to what
might be called academic bibliographic scholarship, on the one hand, and to
the special-library movement with its interest in the special information
needs of business and industry (which I neglect in this paper [see Johns,
1968]), on the other; it also led to the creation of what was called library
science.
Appalled by the absence of rigorous research useful to libraries, C. C.
Williamson in 1930 asked “Can it be that there are no problems in library
service that call for scientific research? Nothing more to learn? No unsolved
problems?” (Williamson, 1931, p. 9.) And he quoted and slightly modified
John Dewey in describing how a library science might emerge:
In effect, the idea of a library science dates from the late 1920s and
specifically from the founding of the Graduate Library School at the Univer¬
sity of Chicago. This school was established in order to bring the best stan¬
dards of higher education to bear on the education of librarians and, through
the introduction of the PhD degree and the work of the school’s faculty, to
promote research in the field. In a statement of policy for the school, Doug¬
las Waples listed what the faculty considered to be the distinctive functions
of the school:
1. The most important single responsibility of the school is to meet the stan¬
dards of scholarship and research maintained by other graduate departments of
the university, both in the character of work undertaken by the staff and by the
research interests of its graduates. 2. The major aim is research, defined as
“extending the existing body of factual knowledge concerning the values and
procedures of libraries and their many aspects, and including the development
of methods of investigation whereby significant data are obtained, tested, and
applied. . . .” 6. Not all of the studies undertaken by the school need be
confined to research in its restricted meaning of “search for abstract princi¬
ples.” In many instances, they may more properly be called service studies,
studies intended to increase the effectiveness of library service. ... 9. An
important function of the school is the preparation, collection, and publication
of monographs whereby the results of significant studies are made available to
the library professions. (Waples, 1931, pp. 26-27.)
350 W. BOYD RAYWARD
31 owe much of this information to the unpublished work of Irene Farkas-Conn, a doctoral
student in the University of Chicago Graduate Library School.
352 W. BOYD RAYWARD
professional and scholarly force only with the publication of its journal
American Documentation in 1950 and with the institute’s transformation
into a general-membership organization in 1953. Nevertheless, the early
emphasis on technology, scientific communication and publication, and new
services offered through new kinds of institutions, anticipated later develop¬
ments within the institute. Here, essentially, was an attempt, stimulated by
technological innovation, to place library and library-related problems in the
context of scientific communication.
Vernon Tate’s editorial description of the scope of American Documenta¬
tion makes it clear that he, like Davis, saw the European documentation
movement as at last crossing the Atlantic to the United States and, from his
point of view, finding in the journal a vehicle that would ensure its subse¬
quent development. [Tate, 1950.] The journal was to deal with “The totality
of documentation,” and the definition of documentation given was that
which had been accepted, Tate noted, by the International Federation for
Documentation,4 for which the American Documentation Institute had be¬
come the American affiliate. “The term documentation refers to the crea¬
tion, transmission, collection, classification, and use of ‘documents’; docu¬
ments may be broadly defined as recorded knowledge in any format.” Tate
also went on to observe of the new work being done by documentalists:
“Traditional methods of communicating and recording knowledge are being
studied, reorganized and supplanted and in some cases revolutionized.”
[Ibid., p. 3.]
In the following decades, it became clear that the documentalists consid¬
ered themselves a breed quite apart from librarians. Unlike librarians, they
were not institution-bound. They belonged to a different environment; they
came to documentation from a wide variety of scientific and technical fields.
They were responding exuberantly to powerful social forces. As Jesse Shera
has said of this period: “First of all interest in the problem of information on
the part of the scientific world and the federal government reached an unprec¬
edented level. Coupled with this was a period of accelerated technological
development.” (Shera and Cleveland, 1977, p. 258.) The exploding world of
scientific literature; new communications and information-processing tech¬
nology; mammoth programs of governmental research calling for new sys¬
tems of information organization, storage, and retrieval—these were the
winds and waves of these new mariners, the documentalists, and the infor¬
mation scientists they formally became in 1968, when the American Docu¬
mentation Institute became the American Society for Information Science.
They invented a new language in which to discuss their work. They became
increasingly involved with the computer and mathematics. Machine-stored,
manipulated, and transmitted data inevitably involved a reassessment of
what constituted a document and what documentation was. The computer
raised the possibility of a wide range of indexing and retrieval experiments
4The new name that The International Institute of Bibliography adopted in 1937.
LIBRARY AND INFORMATION SCIENCES 353
hardly possible before. And, indeed, it may well be that retrieval experi¬
ments in the 1960s constitute a first flowering of information science as a
science.
Certainly, information science as a discipline nourishing certain profes¬
sional activities developed rapidly in the 1960s. In 1965, the first Annual
Review of Information Science and Technology appeared. It is of particular
interest because it represents a relatively early survey of the literature of the
field and its segmentation into what appeared at that time to be the accepted
major disciplinary components. These components vividly exemplified the
importance of the computer in giving shape and direction to both discussion
and research in the field. The following chapters were written explicitly in
relation to the computer: “File Organization and Search Techniques,”
“Automated Language Processing,” “New Hardware Developments,”
“Man-Machine Communication,” “Information-System Applications,” and
“Library Automation.” Two chapters were concerned with aspects of in¬
dexing and abstracting: “Content Analysis, Specification and Control, and
“Evaluation of Indexing Systems.” Another chapter was on “Information
Centers and Services,” central to which was a comparison of work carried
out in special libraries; one was on “Professional Aspects of Information
Science and Technology,” in which were examined problems of terminology
(briefly) and developments in educational programs (at length); and finally,
there was a chapter on “National Information Issues and Trends.”
The change of the name of the American Documentation Institute to the
American Society for Information Science was intended to reflect the chang¬
ing orientation of the society’s members. The adoption of the slippery term
information led to many theoretical and frequently tortuous attempts at
definitions. During the decades of the 1960s and the 1970s, there was also
much discussion about what an ideal curriculum for information science
might look like as opposed to those already in existence for librarianship.
Many of the developments of the late 1960s may be interpreted as the
latest culmination of ongoing attempts to create a more general understand¬
ing of how to facilitate access to recorded knowledge than was possible in
librarianship, which was perceived as being limited to creating collections of
books and related library materials and to managing libraries, or to docu¬
mentation, which focused on a more abstract notion of records of knowledge
or documents. In its turn, information science represented an approach to an
even more abstract level of analysis; it included books and documents while
not being limited to them. Information science has permitted the develop¬
ment of research that is more generally cumulative and generalizable, more
readily enriched by importations from other disciplines. It has led to the
development of new options, procedures, and organizational arrangements
for facilitating access to information recorded, or manipulated, in new ways.
Let us say that by 1970 this was the situation. On the one hand was a
vigorous librarianship, its practitioners still blissful in their relatively undis¬
turbed pragmatism, despite the existence of an important scholarly
354 W. BOYD RAYWARD
apparatus for the profession involving, in part, the nation’s graduate library
schools and despite increasing lip service to the importance of research for
the development of their field. Their journal and monographic literature had
become voluminous, though fragmented. They were organized into a num¬
ber of large, well-established, professional associations with a wide range of
competing, sometimes conflicting, professional functions. On the other hand
was the new, equally vigorous discipline of information science, with a
strong research orientation underlying its professional and service compo¬
nents, its own small fairly homogeneous learned society, a generalized field
of study (whose precise boundaries, however, were in constant dispute), and
a recently established bibliographical and educational apparatus. How did
what was a certain combativeness on the part of information scientists and
librarians change, insofar as it has changed, to mutual accommodation?
DISCIPLINARY CONVERGENCE
It is too simple to say that library schools, aware of the limitations of library
science and recognizing that information science might provide a more re¬
spectable disciplinary base for the profession of librarianship, enticed or co¬
opted information scientists as faculty. Or that it was hoped that if a little
information science were added to their library studies, library-school grad¬
uates might find the job market opening up. Or that information scientists
could not really find a more congenial academic home (or parking lot) in
engineering or computer science or mathematics departments than in library
schools. There may be some truth in all of these statements.
I attribute a disciplinary convergence, insofar as it has occurred, to essen¬
tially two factors. The first was the thoughtful recognition by leaders in the
fields of library science and information science that they were committed to
finding solutions to the same general problems, despite differences in ter¬
minology and orientation. The second was the increasing importance of the
computer in libraries; the need to exploit computers has brought about a
convergence in language and orientation discussed in the next part of this
paper.
In the 1950s, Jesse Shera and Margaret Egan, students of the work of
Bradford and the European documentalists, in a series of still valuable pa¬
pers developed new ideas about librarianship, bibliography, and informa¬
tion. [Shera and Egan, 1953.] They spoke of bibliographic organization
[Shera and Egan, 1951], speculated about classification, sought an intellec¬
tual foundation for librarianship and bibliography in a theory of social epis¬
temology [Egan and Shera, 1951], and tried to understand the significance
and implications of the rapidly changing information technology of the
times. [Shera, 1953.] During his tenure at Western Reserve University and
later, Shera was a witty, thoughtful observer of new applications of the
computer to information handling. He was a librarian eager to reconcile
LIBRARY AND INFORMATION SCIENCES 355
DEVELOPMENTS IN LIBRARIANSHIP
Let us now examine five developments that have taken place in librarianship
during the last several decades. These developments support the idea that,
at least from the librarian’s point of view, there is an emergent disciplinary
integration between librarianship and information science, the historical di¬
vergence and subsequent accommodation of which were discussed in the
preceding section of this paper.
Name Changes
First, and perhaps most superficial, are changes in the way library schools
name themselves. Recently, the Library School at the University of Illinois
changed its name to Graduate School of Library and Information Science.
At Denver, the school became the Graduate School of Librarianship and
Information Management. At Brigham Young University, the school is
called the School of Library and Information Sciences. At Berkeley, it is the
School of Library and Information Studies, while at Chicago, though a range
of information science courses is taught and the school is reported as one of
the first to attempt the systematic integration of a library-and-information
science curriculum [Taylor, 1966], the school continues to be called the
Graduate Library School. Despite these differing appellations, collectively,
we still tend to call all of these schools library schools, but that is no longer
precisely accurate in terms of their interests and scope and how they wish to
present themselves to the rest of academe. The importance of the name
changes lies in what they represent—a shift of emphasis from libraries to
something more general. The curriculum in most of these schools now in¬
cludes courses dealing with computers and computer programming; data¬
base creation, management, and development; information storage and
searching more generally; and behavioral characteristics of various groups
and kinds of information producers and users.
LIBRARY AND INFORMATION SCIENCES 357
Employment Trends
The changes in the names of library schools reflect, in part, another contem¬
porary trend. The graduates of our library schools as well as becoming
librarians now frequently go into information processing or information
management, and they are called information specialists, analysts, or con¬
sultants. This is the second development I wish to mention. To conventional
library jobs, librarians now sometimes add aspects of records management,
special indexing services, media management, and research analysis, and
they often work in what is called an information center. It has become fairly
generally accepted that there is a wide range of professional library and
librarylike work that library-school graduates can perform successfully out¬
side conventional library settings. [Sellen, 1980.]
Journal Proliferation
Terminology
Structure of Associations
A DISCIPLINARY CONTINUUM
which they are maintained, they have begun to attract the interest of librari¬
ans and information scientists.
It is not implausible, either, to think of a museum as a library of artifacts.
Suzanne Briet, writing from the point of view of the European documental-
ists, espouses a definition of documentation almost as broad as this. “An
antelope that has been ‘cataloged’ [in a zoo] is a primary document and other
documents are secondary and derivative.” (Briet, 1951, p. 8.) It seems more
clearly figurative to speak of chromosomes as libraries of genetic informa¬
tion: Certainly neither man nor nature promises generalized access to the
information they contain, despite the advances that have been made in
genetic engineering.
But it is at this point of seeking appropriate definitions for libraries, docu¬
mentation, and information science and of exploring their ramifications that
we may suggest the continued relevance of the historical development ex¬
plored earlier by asking the question: What are the phenomena that informa¬
tion science studies? The preceding discussion has focused on a changing
understanding of the scope and limits of a professional occupation that at¬
tempts to provide information services through relatively permanent and
formal structures. What underlies these changes has been an increasingly
complex and sophisticated view of what constituted recorded information.
First came books; they are documents, and documents contain information,
and access to information is provided for by the systematic manipulation of
the contents of documents and books. Despite all the unknowns, such rela¬
tions are relatively clear, and the study of the processes involved and the
larger environment in which they take place are relatively well cir¬
cumscribed.
When, however, the focus becomes not library and information work, but
information science, the clarity, such as it is, is lost. Information becomes a
phenomenon for scientific investigation, and the contexts within which it is
examined and the investigative techniques to be employed depend very
largely on how information is defined. Information as electrical impulses in
the brain or as represented by certain biochemical phenomena or as the
content of disciplines—this set of different meanings raises a serious prob¬
lem of disciplinary encroachment. When is information science information
science and not epistemology, psychology, biopsychology, physiology or
physics or journalism or the sociology of groups or anthropology?
Thus, an ultimate consequence of the process of seeking an ever more
general perspective for information science, for the problems of access to
recorded and potentially consultable or public information underlying the
creation and development of libraries, may be the loss of all sense of disci¬
plinary coherence in the field. We might argue from the point of view of the
continuum I described earlier that at one end is the librarian constructing a
catalog of books; at the other, because of the way in which information can
now be conceived, is the genetic engineer consulting the information re¬
corded in and then restructuring DNA. Perhaps, however, there is a discon-
LIBRARY AND INFORMATION SCIENCES 363
Ray ward asserts that “we may date the beginning of . . . modern librari-
anship ... in the English-speaking world from the foundation of the Ameri¬
can Library Association (ALA) in 1876 and the Library Association (of the
United Kingdom) in 1877. . . . But within a generation, it became clear that
neither libraries nor the existing organizations of librarians were able to
initiate or even contribute usefully to the study of solutions to certain gen¬
eral bibliographic problems that lay outside the walls of these individual
libraries. This was as true in the United Kingdom as in the United States.”
This is an almost Hegelian interpretation of library history. Lor the Amer¬
ican Library Association to be formed in 1876 meant the coming together of
a large number of like-minded professionals. What had Charles Jewett,
Charles Cutter, Justin Winsor, and Melvil Dewey been doing in the years
before 1876? What had Edward Edwards, Antonio Panizzi, and Andrea
Crestadoro been doing in England before 1877? Britain had passed a Public
Libraries Act nearly thirty years before. The years 1876/1877 do not mark
the beginning of librarianship; they mark its formalization in two national
associations after at least a generation of practice. The international confer¬
ence of 1877 was a declaration of the existence of a solid body of knowledge
and experience.
Jewett had already laid out in the 1840s the three factors essential for
solving the problem of universal bibliography: access to material, reproduc¬
tion technology, and consistent descriptive practice. [Jewett, 1851.] Jewett,
and later, Dewey and Cutter, tried to solve the problem from within the
library world. Ray ward ignores the very real efforts made toward the stan¬
dardization of bibliographic description by librarians like Panizzi, Jewett,
Cutter, and others.
It is true that conventional libraries did not produce a solution to general
bibliographic problems in the last quarter of the nineteenth century—nor did
bibliographers. The reason was not a failure due to any one part of our
general profession, but the effect of historical development—Jewett’s three
problems were never all solved satisfactorily at the same time.
Indeed, as the bibliographers delved even more into the scholarly
minutiae of the differences between copies of early printed books, they
inhibited the production of large-scale scholarly catalogs. It fell to the very
different documentalists (sometimes called scientific bibliographers, to add
to the confusion) to try to collect and organize all the citations in the
scientific literature.
Bibliographic control has never been a central interest of classical bibliog¬
raphy. Classical bibliographers are concerned with differences between
copies of a published work, whereas librarians are concerned with
similarities in order to collect them into subject groups on shelves and into
author/title and subject groups in catalogs and indexes. Documentalists have
always been concerned with specific detail but have customarily used
librarylike techniques to organize their material. They are on the other side of
the librarian from the bibliographer but perhaps not so far away.
368 DAVID BATTY AND TONI CARBO BEARMAN
In the view of Otlet and his colleagues, though general bibliographic organiza¬
tion and control—what they tended to designate as the new field of documenta¬
tion—encompassed traditional librarianship, it went far beyond it. Tradi¬
tionally, libraries were concerned with only a portion of the documentary
record; for the documentalists, the whole of the record lay within their scope.
(Rayward in his paper in this volume.)
Manfred Kochen
1 According to Myron Tribus in his contribution to this volume, Shannon had serious misgivings
about using such an overworked term and did so against his better judgment.
LIBRARY SCIENCE AND INFORMATION SCIENCE 373
2The system was at first called WISE [Kochen, 1972] and elaborated subsequently at a Sym¬
posium of the American Association for the Advancement of Science. [Kochen, 1975.] WISE
was changed to WISDOM at the suggestion of C. T. Larson, who has been working toward
various concrete versions of this idea, such as networks of centers for monitoring and assessing
the state of the built environment, exchanging data, ideas, reports of exemplary practices and
research results in a coherent, integrated form that will lead to improvement, based in part on
his 1953 work with Lonberg-Holms on The Development Index.
374 MANFRED KOCHEN
coupling). Such techniques apply not only to documents that comprise the
written record but also to the individuals who use knowledge as well as to
the fields of knowledge themselves that are only partially embodied in the
record.
Librarianship, library science, documentation, bibliography, and infor¬
mation science in the narrow sense used by Ray ward, have in common a
focus on the written record and the physical documents—books, journals,
manuscripts, musical scores, prints, and so forth—that it comprises. Librar¬
ians are concerned with organizing collections of such documents and
facilitating their use by library patrons. As such, they need to cope with at
least three tasks: (1) continually selecting from the growing record what to
collect (possibly weeding little-used items to make room if there are space
limitations); (2) bibliographic control or organizing and maintaining the col¬
lection as well as maintaining the tools enabling them and their patrons to
determine at any time what is in the collection and where it is; and (3)
reference, to guide patrons in their use of records in the library’s collection.
Information science, in the broader sense in which I interpret it, is, in
contrast, concerned with information, knowledge, understanding, and wis¬
dom. Here, information is used in its technical, scientific sense to denote
what is transmitted over a communication channel to remove a receiver’s
uncertainty about an ensemble. It is both a flow and a pattern that flows over
a channel. This channel may have a memory that stores the encoded infor¬
mation pattern of bits as meaningless codes. The channel may contain a
transducer and transformer that rearrange, encode, decode, or otherwise
process patterns of bits, as in a computer. As a flow, information is a tempo¬
ral change in a pattern. To be informed is to experience a change in some
cognitive structure; to inform is to effect such an experience.
Information becomes knowledge when it is given meaning, usually by a
human mind. A person (or machine) may be said to know the answer to a
previously unencountered question that is not merely a syntatic transforma¬
tion of the statement through which that knowledge was acquired. Knowing
something requires the existence of a knowledge space containing niches for
concepts, ideas, facts, and so forth, that are acquired. These niches are
organized into a web, and the way a niche is embedded into this structure,
relating it directly to many other niches and indirectly to many more, is what
gives it meaning at any time. Knowledge is stock; it may accumulate, but
mere accumulation of knowledge is neither understanding nor wisdom. It is
the latter that can help us to survive and learn to improve our lives.
Knowledge becomes understanding when it contributes to changes in a
knowledge space, adding new connections among niches and new niches,
perhaps revising the entire or large parts of the structure. Certain new items
of knowledge enable the knower to think of and ask questions never thought
of before; they lead to awareness that not all is known and important new
knowledge is needed. Understanding leads to deeper answers to questions
LIBRARY SCIENCE AND INFORMATION SCIENCE 375
regarding how and why, while knowledge is mostly know-what and know¬
how.
Wisdom, in one sense the ultimate purpose of our intellect and the
scientific enterprise, and therefore of central concern to information disci¬
plines, means bringing knowledge and understanding to bear on shaping our
world for human ends by human means. Wisdom governs communication of
information for purposes of control. It includes know-when, that is, the
choice of strategic times for action based on or justified by knowledge and
understanding.
Clearly, this emphasis is far from concern with access to the written
record and even further from understanding such institutions as libraries,
media centers, and so forth. That understanding is certainly necessary but it
is far from sufficient as a professional and scientific foundation for the
knowledge industries. The convergence hoped for by Ray ward is certainly
desirable, but it is not feasible so long as schools of library science take an
attitude of the kind expressed in the following quotation from a resolution by
the faculty of a school of library science:
Jesse H. Shera
Some parts of this paper were used in an article entitled “Information Science and the Theory of
Librarianship,” published in the Indian journal, International Information and Communica¬
tion, vol. 1 (March 1982), pp. 3-15.
379
380 JESSE H. SHERA
which had research resources far beyond those of librarians. The line of
demarcation between documentalists and librarians was even then beginning
to be drawn, and there were some suggestions that library schools should
change their names to schools of documentation. Librarianship was becom¬
ing a tarnished word, and many wanted to dissociate themselves from it as
much as possible.
This movement was intensified by the influence of Claude Shannon and
Warren Weaver and their work at the Bell Telephone Laboratories on the
communication of information. The term information was misleading, but it
was used by members of the American Documentation Institute as being
descriptive of their major concern. Again, the influence of Shannon and
Weaver became prominent in attracting mathematicians to their field, creat¬
ing such technologies as bibliometrics and citation analysis. These phenom¬
ena had long been known to librarians, but their rebirth gave the phenomena
an air of novelty and innovation.
By the late 1960s, the term documentation had become old hat, and the
ADI changed its name to the American Society for Information Science.
Today, even the Special Libraries Association, founded in 1909 by John
Cotton Dana (librarian of the Newark, New Jersey Public Library) to serve
the bibliographic needs of business and industry, is contemplating a change
of name to something akin to an association of information managers. The
flight from library goes on, leaving the old and respected name to typify, in
the main, public library service. The trend would make an interesting study
in linguistic deterioration. It is quite evident that information science is
rooted in attempts to extend the boundaries of library technology and give it
revived respectability by endowing it with a unique name.
The American Society for Information Science and its preceding organi¬
zation, the American Documentation Institute, could provide an excellent
subject for the study of that peculiar amalgam that is librarianship-
information science. Unfortunately, a thorough inquiry of this kind has not
yet been made. However, in 1979, Donald W. King and two graduate stu¬
dents at the library school at Drexel University interrogated the entire mem¬
bership of ASIS in an attempt to provide a profile of the membership of that
organization. [King et al., 1980, pp. 9-17.] The study does provide some
interesting insights. Of the 3894 regular members of the association, 52 per
cent responded. Of the respondents, 24.2 per cent, the largest single group,
were engaged in library administration. The second largest group was 20.3
per cent, composed of managers, directors, or coordinators. Information
technologists, analysts, and scientists came third, with 18.6 per cent, at
which point the percentages drop off sharply. Unfortunately, comparable
data are not available for the early years of the American Documentation
Institute, but I can testify from my own knowledge that the configuration of
the membership was quite different. Most of us would have characterized
ourselves, without hesitation, as librarians.
It is also interesting to note that in the survey of the primary employers of
ASIS members, 35.6 per cent were from industry, 26.9 per cent were from
382 JESSE H. SHERA
colleges or universities, and 12.1 per cent were from the federal government.
In the sample, 56 per cent were women, while in the earlier ADI years, there
was no more than a handful of women. Whether or not these statistics reflect
the actual situation is problematical in that women may be more responsive
to questionnaires than are men, but it does indicate that the new technology
is increasingly of interest to women.
In an editorial on the study, Donald King points out that ASIS members
are concerned with four primary professional areas: operational information
functions, including searching, preparing, and analyzing data on behalf of
others—1490 ASIS members; management of information operations, pro¬
grams, services, or databases—1160 members; information science and
technology, including research and development, systems analysis, and sys¬
tems design—530 members; and educating or training information profes¬
sionals or other workers—360 members. [King, 1980.] We somewhat glibly
assert that librarianship is a field of practice whereas information science is
an area of research—and it was in the early days. But today, if the member¬
ship of ASIS is indicative, information science is approaching an area of
service not unlike that of librarianship. Both activities are concerned with
the transfer (I prefer the term communication) of certain types of informa¬
tion (and again, I prefer knowledge). Information science, however, gets
bigger headlines because of its use of engineering and mathematics in its
technology.
I have indulged myself in this recapitulation of Ray ward’s historical sum¬
mary of the origins of information science in part to amplify some of the
points he makes and also to set the stage, so to speak, for some of my own
argument.
In his summation, Ray ward evinces a certain doubt in his own mind about
an organic interdisciplinarity between librarianship and information science.
Thus, he writes: “When, however, the focus becomes not library and infor¬
mation work, but information science, the clarity, such as it is, is lost . . .
[Immediately there arises the] problem of disciplinary encroachment. When
is information science information science and not epistemology, biopsy¬
chology, physiology or physics or journalism or the sociology of groups or
anthropology?” When he sees information science broadly defined, he dis¬
covers a loss of all disciplinary coherence. He goes on to see this library/
information science relation as a continuum with
at one end ... the librarian constructing a catalog of books; at the other,
because of the way in which information can now be conceived, . . . the
genetic engineer consulting the information recorded in and then restructuring
DNA. Perhaps, however, there is a discontinuity between the two involving
notions of symbolic representation and public consultation that more careful
analysis and definition would reveal.
The point here is not so much the difficulty of definition or the slipperiness
of terminology as the implication that the generalizing process described earlier
L1BRARIANSHIP AND INFORMATION SCIENCE 383
in this paper may still be at work. It is a source of innovation and strength, but
it blurs traditional distinctions and can unsettle professional convictions. (Ray-
ward in his lead paper in this volume.)
ing “control artists,” technicians and mechanics who tinker with informa¬
tion systems but are unable to solve the problems of access and retrieval
because they cannot control ideas. These control artists constantly mistake
the symbol for reality and believe that counting or figuring with these sym¬
bols solves the bibliographic access problem. [Wright, 1981.] This confusion
between data and ideas leads to the confusion between data systems and
idea systems. Our so-called information systems are actually nothing but
data systems. Thus, Wright is close to what Robert Fairthorne has called
“marking and parking” in document retrieval rather than information or
knowledge retrieval.
Librarians have become so concerned with process that they have con¬
fused substance with instrumentation. Processing data can be performed by
machine, but only the human mind can process knowledge or even informa¬
tion. Information science is based on data and their manipulation, not on
ideas.
Science deals with things, things that can be measured, weighed, poured,
fastened, or mixed together, whereas librarians are dealing only incidentally
with things but primarily with ideas, concepts, and thoughts. Librarians are,
or should be, characterized by their knowledge, not by their instrumenta¬
tion. We conventionally say that the computer is here to stay and we must
learn to adapt to its powers and capabilities. I submit that the situation is
quite the reverse: The computer is here to stay, therefore it must be made to
adapt to the librarians’ problems and needs. It must be kept in its proper
place as a tool and a slave, or we will become sorcerer’s apprentices, with
data data everywhere and not a thought to think.
Librarians, ever sensitive to the somewhat meager social esteem ac¬
corded their activities, devised the term library science and eagerly seized
information science as potential supports to their claims to professionalism.
But science does not a profession make, and an overlay of scientific opera¬
tions is not a sine qua non for professionalism. The term profession derives,
of course, from the root verb to profess, to believe in, and etymologically, a
professor is one who acts according to his or her beliefs or convictions.
Wright has recently reviewed the development of the concept from the
Greek arete, which meant skill or proficiency in a particular job and a knowl¬
edge of the job in hand, to the well-known characteristics of a profession as
enunciated by Abraham Flexner. [Flexner, 1915.] Wright cites Socrates to
the effect that if we wish to know what a good shoemaker is and wish to be a
good shoemaker, the first essential is to know what a shoe is and what the
purposes of the shoe are. Only then can we decide what materials are best
for making shoes and what tools should be employed and the best techniques
and methods for using these materials and tools. But such skills and tech¬
niques cannot be successful without an understanding of what is to be pro¬
duced and what functions it is supposed to perform. Wright quite properly
sees a direct connection between knowledge of a task and its performance,
citing the so-called Socratic paradox that “virtue is knowledge,” which was
LIBRARIANSHIP AND INFORMATION SCIENCE 385
‘Stephen J. Tonsor, letter in Wall Street Journal, August 31, 1981, p. 13.
LIBRARIANSHIP AND INFORMATION SCIENCE 387
will soon become like those mythical people of many centuries ago who
mistook for reality the passing shadows reflected on the walls of a cave. So I
conclude with the words of that distinguished librarian of Stanford and sub¬
sequently dean of the library school at Berkeley, Ray Swank: “The long
experience and tested values of librarianship are still our most reliable and
sophisticated resources for the solution of information-handling problems in
most contexts whatever additional help may now be needed.” (Swank,
1981.)
BIBLIOGRAPHICAL R&D
Patrick Wilson
389
390 PATRICK WILSON
First Category
The first and most successful category includes work aimed at improving the
means of storage, manipulation, transmission, and display of bibliographical
information, based on the application of computer technology. Recent years
have seen the development of important new products, systems, and ser¬
vices, such as computer-based services for literature search, computer-
produced citation indexes, and computer-based support systems for catalog¬
ing. While it is impossible to draw a sharp line between research and
development work, it seems clear enough that the work done by the R&D
group leading to such new products, systems, and services has been pre¬
dominantly development work. It is work for the designer of computer sys¬
tems and the programmer. It is technology, not science; the goal has been,
not to produce true statements about the world, but to produce better de¬
vices and systems offering new or better services—product and process
innovations. There are plenty of published documents describing the work,
but the documents are incidental. The literature resulting from this work,
BIBLIOGRAPHICAL R&D 391
Second Category
The second category includes work aimed at improving techniques for creat¬
ing the bibliographical information to be manipulated, stored, transferred,
and displayed. The improved products and processes already mentioned are
advances in operation on given bibliographical information. Attempts to
improve processes of originating the information have been strenuous but
much less successful. At the technical heart of the job of providing intellec¬
tual access to publications are the operations of bibliographical description
(description of everything but the content of a publication) and content
representation through indexing, abstracting, and classification. Work on
problems of bibliographical description has been essentially organizational
or managerial: trying to get agreement on international standards for de¬
scription and trying to reduce unnecessary duplication of effort. Work on
problems of content representation has been theoretical and experimental
rather than managerial: trying to find new principles and radically different
techniques for the analysis and representation of intellectual content. At¬
tempts first to evaluate and then to improve on the performance of tradi¬
tional methods of content representation have been frustratingly difficult and
on the whole disappointing. Experimental attempts at computer-based auto¬
matic indexing and classification have so far led to some advances in under¬
standing but few practical applications. (Computer-produced indexes where
words in titles and abstracts are taken as content indicators without further
analysis are simply ways of avoiding the problem of content representation.)
Abstract analysis of indexing, employing probability and utility theory, has
resulted in formulations of normative conditions for optimal indexing, but
not in altered practices. Improvement in techniques for the representation of
intellectual content is fundamentally different in character from, and enor¬
mously more difficult than, improvement in manipulating strings of charac¬
ters after it is decided what the strings are to be. Strings of characters can be
manipulated, transmitted, and displayed without any attention whatever to
the meanings they convey; producing a correct representation of semantic
content essentially requires understanding the content at some level. Index¬
ing and classification still depend on the rapid intuitive grasp of the scope of
the intellectual content of a text. Numerous novel indexing and classification
schemes for displaying the results of that intuitive grasp of content have
been devised, and many of them put into effect, but it is still true that the
subject indexing and classification schemes most widely used by libraries are
those proposed, and already well developed, a hundred years ago. It is
commonly agreed that these old schemes do not work very well, but improv-
392
PATRICK WILSON
Third Category
The third kind of work done by the bibliographical R&D group is not norma¬
tive but descriptive, devoted to studying some of the characteristics of the
literature that constitutes the input to the bibliographical sector and the use
of that literature. Much of this is quantitative in character, consisting of
bibliometric studies and studies using citation analysis; for example, studies
of the size and growth rates of bodies of literature, concentration and dis¬
persal of literature of various subjects in a population of serial publications,
and obsolescence or decline in use of literature over time. Bibliometrics and
citation analysis are relatively straightforward unobtrusive measures of
some aspects of human communication behavior. [Webb et al., 1966.] Other
quantitative but obtrusive studies of communication behavior are based on
observation, questionnaire, or interview; for instance, the number of hours
spent reading, sources from which people learn of publications useful to
them, the use made of library catalogs, and so on.
There are those who think, or once thought, that such work would lead to
the discovery of laws governing information behavior, such as “laws ac¬
cording to which different kinds of information lose their meaning, validity,
relevance, or value” (Zunde and Gehl, 1979, p. 69); or laws governing the
growth of knowledge; or laws that would stand to information retrieval as
physics stands to engineering. [Weiss, 1977, p. 2.] I take it that we can safely
say that no laws have been discovered: interesting empirical regularities, yes
(Bradford s Law of Scattering being the most prominent example) but laws,
no. [Bookstein, 1976.] A natural law expresses a constraint, a natural neces¬
sity; it tells us what in some sense (hard to clarify precisely) must happen,
and what cannot happen. Simple statements of empirical regularity do not do
that. And it is no more likely that information laws will be discovered by
looking at observable features of documents and uses of documents than
that sociologists or political scientists will discover laws of social and polit¬
ical behavior by looking at what people say and do. [P. Wilson, 1980.] It is
more reasonable, if much more modest, to look on these studies of docu¬
ments and communication behavior as a form of social intelligence, useful to
those concerned with the workings of the bibliographical sector for the
enlightenment they give, if for no other reason. Social intelligence is unlikely
to be directly applicable in the development of new products or processes,
but it can, in the form of market research, indicate the likely reception of a
new service. It can help identify lacunae in services, suggest places where
practical improvements are desirable, give at least rough guidance in the
formation of plans and policies, and spot social changes that are likely to
BIBLIOGRAPHICAL R&D 393
have an impact on the workings of the sector. Research of the sort indicated
here can, of course, be undertaken for its own sake, simply because it is
interesting. There is, in that case, no external criterion of significance, one
fact being as good as another if found to be as interesting. If undertaken to
provide social intelligence to the bibliographical sector, however, research
has an approximate criterion of importance, namely, how much difference it
makes to our plans or expectations or simple understanding of the situation
in which we work. On that criterion, it would have to be admitted that much
of the work actually done is of little or no significance, but that is hardly
unique to the work of this group.
The gathering of social intelligence quite clearly overlaps with work done
by other researchers unconnected with, and only incidentally interested in,
the bibliographical sector. Both in techniques used and in phenomena inves¬
tigated, this work is similar to work done by some sociologists and com¬
munications researchers. More generally, it can be seen as belonging to the
sociology of knowledge. The sociology of knowledge was for a long time too
closely identified with the work of Karl Mannheim to be very productive;
more recently its interests have widened, and a good deal of sociological
inquiry is directed at questions that are also of interest to those gathering
social intelligence for the bibliographical sector. [Holzner and Marx, 1979;
T. D. Wilson, 1981.] Of course, sociologists are not the only social scientists
interested in information and its producers and consumers. How could there
be a branch of the study of social life that did not, sooner or later, have to
investigate information and communication? And bibliographical R&D
might overlap with work in any of the social sciences; as time goes on, we
would expect more overlapping, not less.
The last three categories can be very briefly described as consisting of stud¬
ies of the bibliographical sector itself as it has been, as it now is, and as it
ought to be. Actually, the fourth category, which includes historical studies,
contains a great deal of work consisting of the study of old books as physical
objects, but it does also include studies of the history of libraries, publishing,
and bibliography. The fifth category includes analysis and description of
present conditions in the bibliographical sector, much of which is simply
social reporting, useful since few people are in positions to know much at
first hand about the actual workings of the sector as a whole. Outsiders, or
occasional temporary members of the bibliographical R&D group, have
made important contributions to this category, notably of economic analy¬
sis. [For example, Machlup and Leeson, 1978-1980; Baumol and Marcus,
1973.] The sixth and final category includes studies aimed at improving the
organization and management of the sector (apart from its central biblio¬
graphical components) and analysis of alternative policies for service. The
394 PATRICK WILSON
W. Boyd Rayward
Batty and Bearman are troubled by my restricted palette and small canvas,
and they sketch in some of the historical detail I neglect. Elsewhere, how¬
ever, I have discussed at some length certain general bibliographical devel¬
opments in Great Britain through the latter part of the nineteenth centui y
[Rayward, 1967a], the work of a number of individual bibliographers
[Rayward, 1977a; 1981 b], of the Royal Society for Scientific Bibliography
[Rayward, 1976], and a number of relevant international trends, to which
399
400 W. BOYD RAYWARD
MANFRED KOCHEN
the United States. And, of course, the paths of Otlet and Wells (and Watson
Davis) intersected on at least one notable occasion: The World Congress on
Universal Documentation held under the auspices of the Institute for Intel¬
lectual Cooperation in Paris in 1937.
But no doubt, Leibniz was a forerunner; in the last third of the seven¬
teenth century, he was struggling to find a universal characteristic, devise a
calculus of reason, and create an encyclopedia that would “be the great
instrument of bringing civilization to its highest powers.” (McRae, 1957, p.
38.) In this encyclopedia were to be unified all the arts and sciences; the
encyclopedia would not be merely ‘‘the elaboration of an ordered inventory
of various intellectual possessions. It was absolutely essential for the radical
reform and advancement of the sciences.” (Ibid. p. 42.) Similar to the world
brain and Leibniz’s encyclopedia was the notion of an Office of Publick
Addresse that animated a group of men of letters in mid-seventeenth-century
London at the time of the Civil War and the Commonwealth. How at last to
create Bacon’s House of Solomon; how to mobilize the intelligence of the
day, what we now generically call information, for purposes of social
amelioration; how to coordinate, unify, and allow for the orderly progres¬
sion of the sciences; above all, how to bring about the city of God on earth—
these were among the many religious, scientific, political, and social matters
that Samuel Hartlib, William Petty, and John Drury, among the forerunners
and founders of the Royal Society, wrote about and discussed among them¬
selves and with the great Comenius, visiting London from Bohemia. The
many functions of the Office of Publick Addresse that they variously de¬
scribed were a fascinating combination of a modern employment bureau, the
yellow pages of the telephone directory, classified advertisements in news¬
papers, the community information service offered today by many public
libraries, an information analysis center, a research library, a publishing
house, and a residential institute of advanced study. [Turnbull, 1947; Web¬
ster, 1970.]
Kochen’s WISE and WISDOM schemes are related to a long tradition of
utopian concern for the unification and mobilization of knowledge, of specu¬
lation about ideal forms and uses of encyclopedia. Kochen, however, has
turned away, it seems—with what comes across as dismissive impatience—
from recorded or public knowledge and from systems to make it accessible.
Unlike Leibniz and that group of energetic, enlightened men of the English
Commonwealth, unlike Otlet and Wells, his concern is not with what is
known and publicly available in a readable format in books, scientific pa¬
pers, documents, or even by extension on magnetic tapes, not with the
objective information they contain. His interest is aroused by something
more abstract and subjective, a particular, hitherto elusive quality of the
knower—wisdom. Moreover, he deals with the idea of wisdom in no tradi¬
tional way; his is not the Socratic ideal. For him, wisdom is the product of a
system. WISDOM is a system that will have the capability of making men,
presumably with or without their assent, all knowing and wise, within, we
402 W. BOYD RAYWARD
JESSE SHERA
Shera provides a welcome relief from this kind of thinking. His comments,
graceful and penetrating, amplify the historical perspective I attempted to
provide and offer some interesting data about the composition of the mem-
LIBRARIANSHIP AND INFORMATION RESEARCH 403
bership of the American Society for Information Science. Shera’s chief con¬
cern is to assert the importance of libraries in just those areas that Kochen
denies them relevance, at just the point where WISDOM, a kind of deus ex
machina, supervenes. The library, in Shera’s view, is a social agency, an
instrument of social communication; it is also humanistic in that its “charac¬
teristics, modes of access to its resources, uses, and values are humanistic.’’
Shera is as disillusioned with information scientists as Kochen is disdainful
of librarians. Both scholars, it seems to me, adopt a narrow, perhaps stereo¬
typed, certainly prejudicial view, which I had hoped the perspective I at¬
tempted to provide might obviate. The reasons that Shera gives for abandon¬
ing his early hope that information science would provide “an intellectual
and theoretical foundation” for librarianship seem to me not to be particu¬
larly convincing. They are related to the uses to which the computer has
been put. Certainly, Shera seems to agree that the advent of the computer
precipitated the rapid development of information science and was wel¬
comed by many librarians. But he cites gloomily Kaplan’s Law of the Instru¬
ment and further suggests that information scientists are preoccupied with
the “design, production, implementation, and control of the system.
Under their influence, library schools have lost their sense of the historic
and humanistic mission of the library; they are turning out control artists
and “data mechanics.” Insofar as this has been true (and I am not sure that
it has been), it need not continue. Shera is offering more of a warning than an
argument, for he does accept that librarianship as a form of service a
practice—does interrelate with information science as “an area of inquiry,
of research.”
Libraries are concerned with “sociological and psychological phenomena
not with physical objects and processes,” Shera asserts. Their goal is the
growth of knowledge, and a theory of librarianship, he believes, should be
founded on the concept of “symbolic interactionisms.” If Shera’s position is
at all tenable, then as information science moves beyond data systems,
beyond the gadget and the flickering image, from the thing to the idea, from
information to a concern for understanding and knowledge, it approximates
closer and closer Shera’s ideal librarianship. We are presented with the
irony that Kochen’s WISDOM system may be no more than an idealized
description of the library, perhaps the library of the future, as Shera con¬
ceives it. Here, then, in the terms of my paper, is a convergence with a
vengeance.
Shera’s final warning that control of librarianship as a profession is
threatened by information science does not seem to me to be justified. In a
way, Kochen’s and Shera’s papers are complementary in illustrating the
misunderstanding, hostility, even a certain rivalry, that have existed be¬
tween library and information scientists, for reasons that I find difficult to
understand; there seem to be class or status elements of an interesting kind
involved. In any case, these two papers reaffirm my view, speaking from the
standpoint at least of librarians, that broader perspectives are needed to help
404 W. BOYD RAYWARD
librarianship and the enormously wide variety of libraries that now exist to
adapt, change, grow, even further diversify. Such perspectives will help
them in experimenting with, and in incorporating, new kinds of expertise
and technology; in taking shape according to new and reevaluated old
philosophies and ideals of service; and in drawing on a more developed kind
of scholarship. Let all of these come from where they may.
PATRICK WILSON
cussed. Note that save for the engineers and technicians (and then presum¬
ably not all of them), those whom he mentions in that wildly heterogenous
group would probably describe themselves as librarians or information sci¬
entists. In schools of library and information science (as well as outside of
them, of course) work is regularly conducted in all of the areas Wilson
identifies. But is there more? Wilson speaks of the complex disciplinary
relations members of this group have with others concerned with informa¬
tion and of what may and may not be hoped from the latter who are clearly
not members of the bibliographic R&D community; they are those con¬
cerned with more conventionally conceived sciences and engineering. Their
interest in information seems to complicate for Wilson the question of
definition, or perhaps, more accurately, “designation.”
It seems to me that members of the bibliographic R&D community are
principally library and information scientists, that those who work within it,
whatever their origins, have come to be so designated. The literature pro¬
liferating within it, whether reporting research or development, furnishes the
basis for a multifaceted discipline. Many of those working in the community
have affiliated themselves in the ways I have described. In my view, this
community is not so unorganized, so apparently chaotic as Wilson suggests.
Its research component struggles to cumulate more and better data about a
range of constantly evolving problems, refine our understanding of these
problems, and translate what is known into better practice. (This is more
than process and product development.) Wilson raises, but does not solve,
the problem of how library-and-information science, whatever the extent of
its development, interpenetrates and is developed by the bibliographic R&D
community. It is not enough simply to say or imply that the problem does
not exist. The existence of this community—insofar as the disparate ele¬
ments Wilson describes are not so disparate as to make the idea of a commu¬
nity a ludicrous fiction—does not obviate the various concerns I dealt with
in the last part of my paper; it simply provides a different context for them.
The existence of such a community raises questions of definition, historical
issues of growth and change, contemporary problems of disciplinary assimi¬
lations, and so on. There are, too, occupational issues of how the biblio¬
graphic R&D community, and bibliographic R&D itself, are related to pro¬
fessional practice in libraries and other institutional settings. Wilson s is a
provocative paper built on an intriguing notion. For a community to exist, let
its members work on as many different problems, employing as many differ¬
ent methods as they can devise, the presence of certain commonalities is
required: shared beliefs as to how the problems interrelate, some sense of
common goals, however complex and fragmented. Wilson stresses differ¬
ences. The question of similarities is ignored. Maybe Wilson is simply pro¬
viding his own name for what nowadays would be otherwise described as
library-and-information science (and technology)!
SECTION 6
CYBERNETICS
Murray Eden
There are fields of scientific work, . . . which have been explored from the
different sides of pure mathematics, statistics, electrical engineering, and
409
410 MURRAY EDEN
We may criticize Wiener for claiming too much in the way of priority for
himself and his collaborators. Here and there, he cites collateral and earlier
contributors, but he was not writing a history of scientific ideas. The pro¬
gression was natural enough for a mathematician and statistician in that
time. His continuing interest in biology and neurology—almost an avoca¬
tion—led him to the connections that are so obvious nowadays. They may
not have been obvious in the early 1940s when he was meeting with biolo¬
gists and physicians in the Josiah Macy Jr. Foundation meetings.2 3
At various places in his book, Wiener provides statements of immediate
relevance to the question of interdisciplinary points of contact.
If I were to choose a patron saint for cybernetics out of the history of science, I
should have to choose Leibniz. The philosophy of Leibniz centers about two
closely related concepts—that of a universal symbolism and that of a calculus
of reasoning. From these are descended the mathematical notation and the
symbolic logic of the present day. (Wiener, 1948 and 1961, p. 12.)
Automata theory:
The newer study of automata, whether in the metal or in the flesh, is a branch
of communication engineering, and its cardinal notions are those of message,
amount of disturbance or “noise”—a term taken over from the telephone
engineer—quantity of information, coding technique, and so on. (Ibid., p. 42.)
2 Ludwig von Bertalanffy the biologist and inventor of the term general system theory has a
somewhat different perspective. [Bertalanffy, 1968.] “It is true that cybernetics was not without
precursors [in biology]. Cannon’s concept of homeostasis became a cornerstone in these con¬
siderations. Less well-known, detailed feedback models of physiological phenomena had been
elaborated by the German physiologist Richard Wagner in the 1920s, the Swiss Nobel prize
winner W. R. Hess (1941, 1942) and in Erich von Holst’s Reafferenzprinzip. The enormous
popularity of cybernetics in science, technology and general publicity is, of course, due to
Wiener and his proclamation of the Second Industrial Revolution. (Bertalanffy, 1968, p. 16.)
3 Wiener was a bit wide of the mark here. The current progress of automata theory is to a very
large extent austerely mathematical and draws most heavily on mathematical logic and the
theory of computation.
412 MURRAY EDEN
... it became clear to us [N. Wiener and Walter Pitts] that the ultra-rapid
computing machine, depending as it does on consecutive switching devices,
must represent almost an ideal model of the problems arising in the nervous
system. (Ibid., p. 14.)
... the modern ultra-rapid computing machine was in principle an ideal central
nervous system to an apparatus for automatic control . . . .” (Ibid., p. 26.)
Robotics:
To sum up: the many automata of the present age are coupled to the outside
world both for the reception of impressions and for the performance of actions.
They contain sense organs, effectors, and the equivalent of a nervous system
to integrate the transfer of information from the one to the other. They lend
themselves very well to description in physiological terms. It is scarcely a
miracle that they can be subsumed under one theory with the mechanisms of
physiology. (Ibid., p. 43.)
Pattern recognition:
How do we recognize the identity of the features of a man, whether we see him
in profile, in three-quarters face, or in full face? How do we recognize a circle
as a circle, whether it is large or small, near or far; whether, in fact, it is a plane
perpendicular to a line from the eye meeting in the middle, and is seen as a
circle, or has some other orientation, and is seen as an ellipse? (Ibid., p. 133. )
receive a valuation much as von Neumann’s theory would assign it. (Ibid., p.
165.)
Not at all a bad start on the problem, even though Wiener did not foresee
that much more powerful computers (with the help of some programming
ingenuity) would be made to search the move tree to greater depths than two
or three.
To clarify this heterogeneous set of steps, I shall attempt a parsing of the
sequence in Table 1. The last five ideas of Table 1 appear to define cybernet¬
ics according to Wiener. Most important, therefore, are two related insights.
First, that there is an essential unity in the set of problems in communica¬
tion, control, and statistical mechanisms (noisy phenomena), whether they
are to be studied in the machine or in living tissue. Second, that “the com¬
puting machine must represent almost an ideal model of the problems arising
in the nervous system.” We shall return to this second issue.
It is certainly true that the social system is an organization like the individual,
that it is bound together by a system of communication, and that it has a
dynamics in which circular processes of a feedback nature play an important
part. This is true, both in the general fields of anthropology and of sociology
and in the more specific field of economics .... Drs. Gregory Bateson and
Margaret Mead have urged me, in view of the intensely pressing nature of the
sociological and economic problems of the present age of confusion, to devote
a large part of my energies to the discussion of this side of cybernetics.
Much as I sympathize with their sense of the urgency of the situation, and
much as I hope that they and other competent workers will take up problems of
this sort ... I can share neither their feeling that this field has the first claim on
my attention, nor their hopefulness that sufficient progress can be registered in
this direction to have an appreciable therapeutic effect in the present diseases
of society. (Ibid., p. 24.)
neuromuscular
a
u
p
Dh
Explaining
behavior
c
cd
o>
"O
<D
a
Af
>, hn
H
a>
s
a
o
<U
>
2 Control
theory
<D
Q 13
O
-Q
03
<u
Cl
>,
H
J*i
U a>
o >
<L>
o £ C/3
a
u
e cd o 01)
p
Oh
O .cd
c ao ^
c *5 Id <L>
22
c oj C 50
o cd
C/3 att *C <u • 2 x>
03 .S Cd
a f<u 'Zn o 5- 2 *c T3 u
-5 o ?£ <L> <D n> C *2
3 Cd v a
u. .2
D
■4—*
c/d -a Q co c: > u 0-. cd
C/D U3
a <D
<D JO
Q.
>>
H
•o
cd
Glj J2
c 2 ^ m
<D
P <u I u + ^ c
O Q
E "u c p
a D C cd
jo a- *s O c/3
o <N
cr'-S
<L> E ^
C/3
+ o -J3
> o « g > <u
u.
<L>
U z g H O*
Q
414
<u
o
c
1)
<v G
> .o>
G-
c
o
V , ,
Q. o
C/5 <u
D
u fco
o
OD 0)
O X <3
<D
O -g
<u
X) X
C/5
c o
-o U
X
03 g
cd <u
C/5 <D X
o a. ■*“*
‘X a c
o3 as Gn
■■ ■
— O
a £
£
E
■— ■o
G<
C5JJ
o X
3
-a
c
o
C/5
a
c G
X X <u
o ■*—' c
<u a> <D
0) sC/5
£
o
c
c
<D 0
•o c
‘g
C/5
03 <u
*0.
T3 o
C “D E
o3 G
<D C/5
c
§ <u
C/3
X
f-T £ H
o
a -Cl
415
416 MURRAY EDEN
society are not only statistical, but the runs of statistics on which they are
based are excessively short.” He hints at the issues of ergodicity, stationar-
ity, and changes of state:
There is no great use in lumping under one head the economics of steel industry
before and after the introduction of the Bessemer process, nor in comparing
the statistics of rubber production before and after the burgeoning of the auto¬
mobile industry and the cultivation of Hevea in Malaya. Neither is there any
important point in running statistics of the incidence of venereal disease in a
single table which covers both the period before and that after the introduction
of salvarsan, unless for the specific purpose of studying the effectiveness of
this drug. (Ibid., pp. 24-25.)
4 For a discussion of the philosophical aspects of their approach the reader may wish to consider
the comments of Berlinski [1976].
5Nina R. Thompson, ed., Cumulative Book Index 1967-1968 (New York: H. W. Wilson Co.
1971), vol. 1, p. 814.
CYBERNETICS 417
The situation is by no means ideal, and there is a danger that cybernetics will
become generally regarded as an up-to-date form of Black Magic, as a sort of
twentieth century phrenology. If this unfortunately happens, then one whole
field of research, at the border of many disciplines, will be robbed of the
support it needs. (Young, 1969, pp. ix-x)
Fourteen years later, it must be said that the situation has not improved and
that in large measure Young’s melancholy prophecy in the last paragraph of
his preface has been largely fulfilled.
Although I have not reviewed in more than a cursory way Russian litera¬
ture available in English, it may be worth noting the somewhat different
course of cybernetics in the Soviet Union. Wiener, who was a humanist and
something of a socialist in his political leanings, was much disturbed by the
original reception of his work in the Soviet Union. It appears to have been
relegated to the category of bourgeois science and, hence, dismissed. How¬
ever, following the defrocking of Stalin as a scientist, Wiener’s cybernetic
principles were firmly endorsed. In a manner that may be typical within
Russia’s science establishment, cybernetics—almost overnight—became a
fundamental discipline; institutes were created, academic departments es¬
tablished, ajournal was initiated, courses and textbooks were prepared. One
striking difference remained. Western physical scientists and engineers
rarely used cybernetics in titles of papers and referred, instead, to the mathe¬
matical name of the technique being applied to their research: control, time-
series analysis, statistical communication theory, automata theory, and so
forth. Russian literature tended to use cybernetik whenever any of these
techniques were being applied. Science Citation Index for 1980 includes
papers of Russian origin, with such titles as “Making Ultrasonic Inspection
of Welded Joints More Efficient by Means of Cybernetics’’ and “Group
418 MURRAY EDEN
LIVING SYSTEMS
Putting aside the curious story of the linguistic depreciation of the word
cybernetics, what of the ideas behind the term, especially with respect to
mind and the nervous system? Every age compares life with its contempo¬
rary machinery.
Consider how interpreters of mind and nerve attempted to explain these
phenomena. A central concept for Aristotle was pneuma. I draw the follow¬
ing quotations from the biologist, Thomas S. Hall, who has written what he
calls a ‘‘preliminary account” of the “difference between living and nonliv¬
ing things.” (Hall, 1969, p. viii.)
Diodes (ca. 300 B.C.) and Praxagoras (ca. 300 B.C.) had given the psychic
pneuma its source in the heart whose arteries distribute it to the body generally
whereas the veins carry blood. . . .
It remained for Praxagoras’ pupil Herophilus to take the giant stride of identify¬
ing the motor nerves and their role in efferent conduction, as well as the
sensory nerves which he probably regarded as conductors of a sensory
pneuma. The important point to note here is that Herophilus and his successor
at Alexandria, Erasistratus, now returned the brain to the position of central
organ of sensation and volition. Herophilus showed through dissection that
nerves arise from the brain. Erasistratus’s doctrine is already familiar to us.
I chose cycle because the concept of cycles is important within cybernetics and made a
correction for purely chemical terms such as cyclic AMP. Science Citation Index, part 10,
permuterm subject index, columns 14496-14497, ISI Inc., Philadelphia, Pa. [1981],
CYBERNETICS 419
The arteries distribute vital pneuma to the body generally, he said, including
the brain. Here psychic pneuma is formed and travels outward over the nerves.
Galen’s ideas are notable especially because of the extent to which actual
experimental evidence influenced his opinions. (Ibid., pp. 161-162.)
We are tempted to think that the Greeks purified their writings from the
taint of technology much as modern mathematicians delete their marginalia
and numerical examples, which detract from the elegance of their papers.
Nevertheless, there is at least circumstantial evidence of a technological
marker for pneuma. A few centuries after Aristotle, Hero of Alexandria, in
his work Pneumatica, is reported to have given several examples of a kind of
apparatus that may be the prototype of existing engines. Steam or steam-
water-mixtures operated mechanisms that could move temple doors without
human intervention. We may also note that Aristotle’s use of pneuma is pure
metaphor, but experimental findings were beginning to introduce their con¬
straints on the model. “As for method, although Galen’s experiments were
often inconclusive, they were also numerous and relevant and were not to be
matched in ingenuity until the seventeenth century.” (Ibid., p. 163.)
The age-old vitalist/mechanist argument in biology is well known. But we
are not here concerned with the argument per se but rather the role of the
automata or pseudoautomata that have been known in history and myth as
models for mechanistic thought.
Descartes was the most explicit early proponent of the mechanistic view.
He closes his Treatise of Man with the following paragraph:
I desire that you should consider that all the functions that I have attributed to
this machine [he here enumerates the functions] . . . imitate as perfectly as
possible those of a real man; I desire, I say, that you should consider that all
these functions follow quite naturally solely from the arrangement of its parts
neither more nor less than do the movements of a clock or other automaton
from that of its counterweights and wheels; so that it is not necessary to
conceive in it any other principle of movement or of life than blood and spirits
set in motion by the heat of the fire that burns continually in its heart and is of
no other nature than all those fires that are in inanimate bodies. (Descartes in
Le Monde, published posthumously 1664; as quoted by Hall, 1969, p. 262.)
whole organism and its main working parts were viewed as machines. For
Descartes (1635) the whole animal was an automaton and so were such working
components as the heart and the various pairs of reciprocating muscles. Borelli
(1685) had a comparable attitude toward the bird and its wings; the whole bird
was a flying machine of which the wings are necessary working parts. Harvey,
writing earlier than Descartes, compared the propulsion of blood through the
heart to that of water through a water bellows.
Some theorists carried the mechanical analogy even beyond the level of things
potentially visible through a microscope. Thus both Descartes and Giovanni
Borelli . . . extended mechanical models all the way to the unit particle (Hall
1969, pp. 228-229.)
If we have a series of neurons, with a fiber of the first in contact with a fiber of
the second; a fiber of the second in contact with a fiber in the third, and so on;
and if we stimulate the first neuron properly, it will stimulate the second, the
second will stimulate the third, and so on. Obviously, some “process” passes
through the neuron, passing into the cell body through one fiber, and out
through another: because a neuron excited at the tip of one fiber will excite
another neuron with which another of its fibers is in contact. This “process” is
called “nerve current.” Its exact nature is unknown, and is the subject of
various theories, but it is analogous to the burning of the powder in a fuse.
Suppose we lay a number of short pieces of fuse end to end, and light the free
end of the first. The combustion process will run through the first piece, ignite
the second, and so on through the whole line. Yet nothing travels, except the
process of combustion. The chief differences between the action of this line of
fuses, and the action of a chain of neurons are that the neuron has a rapid
recovery, becoming quickly ready for another discharge; and that the neuron,
in many cases, has numerous branches of its fibers in contact with many other
neurons, and stimulates, at a given discharge, some of these without stimulat¬
ing the others. (Dunlap, 1922, p. 177.)
only one point. Current emerging from the cord or brain stem at a given point
can go to only one effector or effector group. The central neurons, however,
provide a multiple switching system, by which an afferent route can be con¬
nected with any one of several efferent routes. If we consider the cord, brain
stem and hemispheres together, they may be compared to the exchange of a
telephone system, through which any calling phone can be connected with any
other phone. The difference is that the “sending” (receptors) and “receiving”
(effectors) stations in the nervous system are distinct; whereas they are com¬
bined in the telephone system. The nervous system is a “one-way system.”
But within this system a given receptor can be connected with almost any
effector, or with several at the same time; and conversely, a given effector may
have connected to it any one of a wide range of receptors or various combina¬
tions of receptors. (Ibid., pp. 183-184.)
This was the state of affairs in neurology until the second half of the twen¬
tieth century. Mechanics was the metaphor, and contact between neurology
and psychology was limited largely to the fact that behavior is the expression
of nervous activity. Wiener, as we know, used the computer as a model.
Wiener was among the first to see the computer analogy, and he certainly
contributed to its popular diffusion.
Let me now come to another point which I believe to merit attention. It has
been clear to me that the modern ultra-rapid computing machine was in princi¬
ple an ideal central nervous system to an apparatus for automatic control; and
that its input and output need not be in the form of numbers or diagrams but
might very well be, respectively, the reading of artificial sense organs, such as
photoelectric cells or thermometers, and the performance of motors or sol¬
enoids. With the aid of strain gauges or similar agencies to read the perfor¬
mance of these motor organs and to report, to “feed back,” to the central
control system as an artificial kinesthetic sense, we are already in a position to
construct artificial machines of almost any degree of elaborateness of perfor¬
mance. (Wiener, 1948 and 1961, pp. 26-27.)
At that time [1943] Mr. Pitts was already thoroughly acquainted with mathe¬
matical logic and neurophysiology, but had not had the chance to make very
many engineering contacts. In particular, he was not acquainted with Dr. Shan-
422 MURRAY EDEN
non’s work, and he had not had much experience of the possibilities of elec¬
tronics. He was very much interested when I showed him examples of modern
vacuum tubes and explained to him that these were ideal means for realizing in
the metal the equivalents of his neuronic circuits and systems. From that time,
it became clear to us that the ultra-rapid computing machine, depending as it
does on consecutive switching devices, must represent almost an ideal model
of the problems arising in the nervous system. The all-or-none character of the
discharge of the neurons is precisely analogous to the single choice made in
determining a digit on the binary scale, which more than one of us had already
contemplated as the most satisfactory basis of computing-machine design. The
synapse is nothing but a mechanism for determining whether a certain combi¬
nation of outputs from other selected elements will or will not act as an ade¬
quate stimulus for the discharge of the next element, and must have its precise
analogue in the computing machine. The problem of interpreting the nature and
varieties of memory in the animal has its parallel in the problem of constructing
artificial memories for the machine. (Wiener, 1948 and 1961, p. 14.)
PSYCHOLOGY
The history of thought upon the psycho-physical problem is in the main the
history of the way in which Animism, the oldest and, in all previous ages, the
most generally accepted answer to it, has been attacked and put more and more
upon the defensive in succeeding centuries, until towards the end of the
nineteenth century it was generally regarded in academic circles as finally
driven from the field. I have therefore given to the historical chapters the form
of a history of Animism. . . .
The modern currency and usage of the word derives chiefly from Prof. Tyler’s
“Primitive Culture,” and I use it with the general connotation given it in that
celebrated treatise. The essential notion, which forms the common foundation
of all varieties of Animism, is that all, or some, of those manifestations of life
and mind which distinguish the living man from the corpse and from inorganic
bodies are due to the operation within him of something which is of a nature
different from that of the body, an animating principle generally, but not neces¬
sarily or always, conceived as an immaterial and individual being or soul.
(Ibid., p. 14.)
7My judgment has been made in light of the publishing history of this popular text. The copy
available to me is a version of the twentieth edition that appeared in a new size in 1963 and was
reprinted in 1967. The twentieth edition was copyrighted in 1949. However, the preface to the
edition is dated February 26, 1947.
424 MURRAY EDEN
But consider the psychology text by Floyd Ruch revised for the ninth edition
in 1967. [Ruch, 1937 and 1967.]
Although psychologists and neurologists still have much to learn about our
brains and nervous systems, they are sure that underlying every thought, every
perception and every action is a pattern of neural activity. As we encounter
stimulus situations of all kinds, messages are received, evaluated, integrated
and stored, and other messages are sent out to the various organs of response.
(Ibid., p. 41.)
new stimulus. Thus we see that our formula will not be really adequate unless
we add the element of sensory feedback. (Ibid.)
During the last decade, teaching of general psychology has been turned
upside down. Peter Lindsay and Donald A. Norman entitle their text Human
Information Processing: An Introduction to Psychology. They begin the
book with perception described in terms of pattern recognition, rules, and
feature analysis. They describe “neural information processing” in terms of
idealizations of the nervous system as logic elements and circuit diagrams.
Feedback control enters even in their discussion of motivation. The ter¬
minology of cybernetics as construed by Wiener is on virtually every page,
but neither Wiener nor cybernetics is mentioned. [Lindsay and Norman,
1972.]
BIOLOGY
The nerve fiber can be compared, from the point of view of its function, with a
cable carrying information. The approach of the communications engineer—
information theory—can also be applied to the nervous system. This body of
knowledge, together with control theory, is often called cybernetics; especially
in Europe, its application in biology is termed biocybernetics, though such
general terms are used less in the United States. (Ibid., p. 68.)
great deal of space is devoted to articles that require the notions of informa¬
tion and pattern recognition. [Held, Leibowitz, Teuber, 1978.] But since we
have already seen them in introductory texts, there is no use belaboring the
point.
Outside the domain of neurology or psychology, we find traces of the
ideas of cybernetics in many corners of biology. Carl F. Rothe contributes a
chapter to Selkurt’s Physiology, which he entitles “Regulations of Visceral
Function—B. Homeostasis and Negative Feedback Control.” [Rothe in Sel-
kurt, 1971.] He presents a block diagram for a negative feedback-control
system but includes no mathematics.8
Brian C. Goodwin’s book deals with a biological theory intended to pro¬
vide insight into the basic formal organization of the highly evolved and
elaborate forms of life. He casts much of his argument in terms of control
models and makes heavy use of the mathematics of control theory and
probability. [Goodwin, 1963.]
David Givol, an immunologist, begins a paper entitled “A Structural
Basis for Molecular Recognition: The Antibody Case” [in Cuatrecasas and
Greaves, 1976] with the sentence “Molecular recognition stands at the roots
of biology. . . . Recognition via combining sites is a property of enzymes,
antibodies, hormones, receptors, lectins and perhaps other classes of pro¬
teins.” (Givol, 1976, p. 3.)9
Earlier in this collection, Melvyn Greaves writes of “biological communi¬
cation”10 as well as “signals and the cell surface.” [Greaves in Cuatrecasas
and Greaves, 1976.]
8 For the nonbiologist, it may be worthwhile to explain the distinction between homeostasis and
feedback-control. The concept of homeostasis is due to Walter B. Cannon. [Cannon, 1929.]
Rothe defines it thus: “Homeostatic mechanisms operate to counteract changes in the internal
environment which are induced either by changes in the (external) environment or by activity of
the individual so as to maintain a constant and optimal internal environment.” Rothe goes on—
alas—to equate homeostasis with negative feedback-control. He is wrong. Feedback models in
the physiology of organisms are instances of homeostasis, but the great bulk of homeostatic and
adaptive models for the explanation of stability of composition and chemical activity in organ¬
isms rarely appeal to the notions of set-point, error function, and negative feedback. Of course,
flow diagrams are put down, and (generally nonlinear) partial differential equations are used
extensively, but the framework is that of physicochemical kinetics rather than control theory.
We may regard homeostasis as the biological equivalent of LeChatelier’s Principle augmented
by the hypothesis that concentrations of material substances and the energetics of biological
internal processes retain their constancy (in the common sense of that word) because the
variation of any one is tied to very many, if not all, of the others.
y Molecular recognition draws on thermodynamics in order to give meaning to the notion of
information. Molecular biologists measure affinity or association constants. The logarithm of
this quantity is proportional to the (Helmholtz) free energy of binding. Very rarely, the entropy
is evaluated, which would bring the concept into formal equivalence to the concepts of physical
information originated by Leo Szilard [1929] and John von Neumann [1932 and 1955],
1 ’The notion of specificity in the action of biologically active substances is at least seventy years
old. Paul Ehrlich formed the concept of a receptor theory from immunological studies on toxin-
CYBERNETICS 427
So much for biology. The basic concepts of Wiener’s cybernetics have be¬
come embedded in virtually every branch of scholarship. The notion that
information can be handled as a mathematical quantity; that this quantity
mediates the transmission of energy and, therefore, can be used to control it;
that statistical uncertainty is itself a quantity related to (and degrading)
information; that information can be used to guide decision-making and
clarify our own descriptions of real-world phenomena; that information may
be transformed by computation; and that information and control can be
combined to create complex mechanisms that behave as if they were intelli¬
gent—these ideas, I maintain, are pervasive.
I feel it necessary to repeat that I am not commenting on the question of
Wiener’s priority in all his claims, nor is it relevant to my argument that the
domain of cybernetics has changed with the times and now tends to exclude
many of the ideas Wiener had included originally. Simply put, it seems to me
that the ideas Wiener set down in his book and which he considered to be
within cybernetics are now the common currency of science. It seems fruit¬
less to attempt to prove my assertion. Mathematics—pure or applied—is
there to be used whenever a scientist who has at least a nodding acquain¬
tance with a particular subset of mathematics decides it is relevant to his
problem. Moreover, because Wiener adjoined so much of mathematics to
communication and control theory, we would be hard put to identify some
area (not even differential topology or algebraic number theory) that could
antitoxin interactions. Even earlier Emile Fischer had attributed enzyme-substrate interactions
to a lock-and-key (schloss und schlusset) mechanism. [Ehrlich, 1956, vol. 2, p. 178.] 1 wonder
whether the introduction of the words recognition and communication helped. The fact remains
that they are now part of the biologic vocabulary. [Cited in Greaves, 1976, p. 15.]
428 MURRAY EDEN
Negative feedback The psychological concept of goal was identified with the
mechanism of setting up a generalized servomechanism to reduce the differ¬
ence between an input goal parameter and an observed system parameter. This
idea was exploited in various mathematical directions, but the secret of intelli¬
gence was not to be found in “optimal control” or similar knowledge-free
theories. Nonetheless, the difference-reduction concept, reformulated in terms
of a symbolic description of differences, finally became a key concept in
artificial intelligence in the General Problem Solver system of Newell and
Simon.
attempting to match image against image, rather than description against de¬
scription. The issues raised in this exploration—how a global or gestalt charac¬
teristic of a situation is discerned from an ensemble of local features—slowly
evolved into a very rich collection of theories of description and representa¬
tion, so rich, in fact, that the subject is still difficult to survey and criticize.
Stochastic learning The secret of creativity was sought in the area of con¬
trolled random search, both in models of learning and in models of problem
solving. The “stochastic learning models” of that era’s mathematical psychol¬
ogy did not lead anywhere, nor did experiments on “programmed evolution"
or “random neural networks.” A somewhat later approach based on “percep-
trons,” which were self-adjusting learning devices, seemed more promising but
subsequently died out. (Minsky, 1979, pp. 401-402.)
Having concluded that the cybernetic hypotheses are part of the subcon¬
scious culture of science and technology, we are led also to the conclusion
that cybernetics does not constitute a well-defined scholarly discipline nor
does it connote a group of professionals with common technical problems
and interests. There remain a few relatively minor journals whose names
concatenate several terms; thus, Journal of Cybernetics and Information
Science; Biological Cybernetics, Biofeedback and Self-Regulation (formerly
published as Kybernetik); Kybernetes: An International Journal of Cyber¬
netics and General Systems; and Transactions of the IEEE Systems, Man
and Cybernetics Society. Occasionally symposia—also with concatenated
titles—are held and their papers published; thus, Progress in Cybernetics
and Systems Research. [Trappl and Hanika, 1975.] There is an International
Association of Cybernetics.11
"There are some differences between terminology in the United States and Europe. The editors
of the symposium just cited are president and vice president, respectively, of the Austrian
Society for Cybernetic Studies. Professor Robert Trappl, the president, is also professor of
biocybemetics and bioinformatics at the University of Vienna Medical School. The Technical
University of Berlin has a cybernetic department (Fachbereich Kybernetik). I also note re¬
search groups on sociocybemetics in the computer science laboratory of Uppsala University
and a Department of Cybernetics at Brunei University, England.
430 MURRAY EDEN
The Russian mathematical scientist Valentin F. Turchin has cast the ideas of
cybernetics into a more formal mathematical way of looking at things than
did Wiener. [Turchin, 1977.] For Turchin, cybernetics is based above all on
the concept of the system, consisting of other objects called subsystems; so
cybernetics considers the relations between systems and subsystems. A
system or subsystem is characterized by its state, a concept that relies on
our intuition of time; thus, cybernetics studies the organization of systems in
space (real and abstract space) and time. Organization in time when it is
purposive is called control. Finally, cybernetics tests the concept of hierar¬
chy; the description at one hierarchical level will be different from that at
another level. Hierarchical structure or subsystems are intended to recog¬
nize or classify concepts derived from lower levels. Hierarchies also estab¬
lish the relations of goals and plans for the system by which recognition and
the consequent control decisions are made.
When cybernetics is cast in this framework, it is easy to understand the
growth in the use of the term systems analysis or systems theory as an
1'The College Blue Book, 18th Edition—Degrees Offered by College and Subject (New York:
Macmillan, 1981).
13 Wiener devotes a significant portion of the pages of Cybernetics to mathematical discussions.
We have the impression that the mathematics was introduced for illustration rather than for
teaching the mathematics or its application.
CYBERNETICS 431
equivalent to cybernetics. Many times the words are conjoined; for example,
the largest society dealing with cybernetic issues per se is the IEEE Sys¬
tems, Man and Cybernetics Society.
Systems theory is a domain of scholarship that exists in uneasy relation to
that of cybernetics. The originator of the term general system theory and one
of its major expositors, Ludwig von Bertalanffy, subsumes cybernetics
within general systems theory. [Bertalanffy, 1968.] As with Wiener, Ber¬
talanffy has no succinct definition for his newly coined phrase. It would
appear that all problems not solvable by classical science are problems of
general systems theory:
As has been said, there are various approaches to deal with such problems. We
intentionally use the somewhat loose expression “approaches” because they
are logically inhomogeneous, represent different conceptual models, mathe¬
matical techniques, general points of view, etc.; they are, however, in accord
in being “systems theories.” (Ibid., p. 19.)
Not only are general aspects and viewpoints alike in different sciences; fre¬
quently we find formally identical or isomorphic laws in different fields. In
many cases, isomorphic laws hold for certain classes or subclasses of “sys¬
tems,” irrespective of the nature of the entities involved. There appear to exist
general system laws which apply to any system of a certain type, irrespective
of the particular properties of the system and of the elements involved.
with general system theory in the title was in proof when it was destroyed in
World War II.14
The systems theorist George Klir holds a somewhat more restricted view
of the general systems theory but includes cybernetics within systems
theory. We are indebted to Klir for explicitly distinguishing between cyber¬
netics as a science and cybernetics as a viewpoint. [Klir, 1970.] I could find
in his work no suggestion that an analogous division of systems theory
would be appropriate.
A currently active general systems theorist, Mihajlo Mesarovic, offers
another definition, one that stresses the mathematical nature of the theory:
[The] conceptual basis for systems theory might be summarized in the follow¬
ing observation: Study of any real-life phenomena reveals two aspects—in¬
formal, dealing with the meaning, interpretations, significance, objectives,
values and the like—and formal, dealing with the form (structure) in which the
relationship between the attributes appears. Systems theory is concerned with
the second, formal, aspect of the observations. Apparently, these formal rela¬
tionships are invariant with respect to the specific nature of the phenomena
under consideration.
Systems theory, then, is based on the following fundamental premises:
(a) A theory of any real life phenomena (biological or otherwise) is always
based on an image, termed a model.
(b) Without introducing any constraints whatsoever the formal, invariant,
aspects of the model can be represented as a mathematical relation. This
relation will be termed a system.
l4The citation for this work appears to be missing from General System Theory, although there
may be an inconsistency between the text (p. 14) and the bibliography. The text suggests the
first paper was to appear in Deutsche Zeitschrift fur Philosophic, while the bibliography, or¬
dered chronologically, cites “Zu einer allgemeinen Systemlehre,” Blatter fur Deutsche
Philosophie 3/4 (1945), extract in Biologia Generalis 19 (1949), 114-129. Are these one and the
same?
CYBERNETICS 433
pret it in terms of the (different) ways in which later workers describe its
domain? Other domains of scholarship do not ordinarily pose such prob¬
lems.
There is a direct line of descent in the history of control theory from Harry
Nyquist through Wiener, followed by such control theorists as Richard Bell¬
man, Rudolph Kalman, and onward to the present day. It is not implausible
to suggest that James Clerk Maxwell would agree with his intellectual de¬
scendants as to the scope of a theory of governors that began with his work
of 1868. [Maxwell, 1868.]15
Very much the same argument can be made with regard to the history of
information or communication theory. Claude Shannon in The Mathemat¬
ical Theory of Communication asked and largely answered a question that
was undoubtedly close to Wiener’s problems with noise in stationary time-
series analysis: “How is an information source to be described mathemat¬
ically and how much information in bits per second is produced in a given
source?” (Shannon and Weaver, 1949, p. 10.) Shannon thought of his work
as an extension of the general theory of communication contained in the
work of Nyquist [1924] and Ralph Hartley [1928] twenty years earlier. In a
comment appended to Shannon’s seminal paper, Warren Weaver acknowl¬
edges the contribution of several precursors.
Dr. Shannon’s work roots back, as von Neumann has pointed out, to
Boltzmann’s observation, in some of his work on statistical physics (1894), that
entropy is related to “missing information,” inasmuch as it is related to the
number of alternatives which remain possible to a physical system after all the
macroscopically observable information concerning it has been recorded. L.
Szilard (Ztsch. f. Phys., Vol. 53, 1925) extended this idea to a general discus¬
sion of information in physics, and von Neumann {Math. Foundation of Quan¬
tum Mechanics, Berlin, 1932, chap. V) treated information in quantum me¬
chanics and particle physics. Dr. Shannon’s work connects more directly with
certain ideas developed some twenty years ago by H. Nyquist and R. V. L.
Hartley, both of the Bell Laboratories; and Dr. Shannon has himself empha¬
sized that communication theory owes a great debt to Professor Norbert
Wiener for much of its basic philosophy. Professor Wiener, on the other hand,
points out that Shannon’s early work on switching and mathematical logic
antedated his own interest in this field; and generously adds that Shannon
certainly deserves credit for independent development of such fundamental
aspects of the theory as the introduction of entropic ideas. Shannon has natu-
15 Maxwell writes: “I propose at present, without entering into any details of mechanism, to
direct the attention of engineers and mathematicians to the dynamical theory of governors.”
(Maxwell, 1868, p. 106.)
434 MURRAY EDEN
For completeness, we note that Shannon also acknowledged the early use by
Ronald A. Fisher of an information measure different from his: the quantity
of information in a sample drawn from a density distribution with finite
expectation. Colin Cherry carried the history back much further, but it need
not concern us here. [Cherry, 1951.]
The appearance of Shannon’s work had an explosive effect, not only in
communications theory, but in sparking a multitude of attempts to apply his
theory to all sorts of processes that appeared to be related to the transmis¬
sion of objects with semantic content. How much of this efflorescence was
due to Shannon and how much to Wiener is impossible to assess; the appear¬
ances of Cybernetics and “The Mathematical Theory of Communication”
(as a paper) were contemporaneous [1948]. However, there are several im¬
portant differences. Shannon presented a developed mathematical theory;
he offered a transparently simple set of measures of information. He avoided
entirely any allusion to the many kinds of transactions that people are likely
to call informative. Rather, he asked how to measure information, and he set
the ground rules for models that could conceivably use his measures.16 Then,
Shannon, in a simple example, showed how to estimate the entropy of a
natural information source, the English language.
Information theory has developed much as any other branch of applied
mathematics. Many applications have been developed for communication
problems. Some of them led to newer, richer theory, and these were applied
in their turn. And so on to the present.
This is not the place to elaborate on general systems theory, but we must
distinguish between systems theory as a viewpoint and mathematical sys¬
tems theory. The latter is quite clearly to be included within mathematics
and continues to be an active area of mathematical development. It is in the
direct line of descent of govenors, servomechanisms, feedback, and optimal
control previously alluded to. Rudolf Kalman, Peter Falb, and Michael Ar-
bib in the introduction to their text Topics in Mathematical System Theory
lay emphasis on “the concepts of states, control, optimization and realiza¬
tion.” [Kalman, Falb, and Arbib, 1969.] Suffice it to say that systems theo¬
rists and cyberneticians occupy the same turf. The choice of a name for a
new domain of scholarship is a sociological rather than a scientific question.
Scholars of the West tend to favor systems theory, the Eastern bloc appears
to prefer cybernetics; there are some who will use both, and others who use
neither, favoring instead more restricted labels to classify their work.
l6This is by no means a trivial exercise. The probability distribution on an infinite set of possible
messages must be properly defined or else the apparatus collapses. Wiener provided no handy
mathematics; although at times he urges caution, he did not provide the apparatus for imposing
restraint.
CYBERNETICS 435
Biology, as was indicated earlier, has adopted the insights of cybernetics and
many of the terms used in control theory, information theory, and com¬
puters. While I have attributed it to Wiener’s influence, it may well have
been due to the rise of systems theory. That is to say, biologists have
presented their ideas in fora of both kinds. The Yale physiologist, Talbot
Waterman, has put forward his view of the importance of systems analysis at
a symposium on Systems Theory and Biology.
But homeostasis itself is truly dynamic in nature despite its proper distinction
from the reference-input-following servo-type control. This is so because the
physiological steady state is neither just a passive resistance to change nor a
mere compliance to pattern imposed mainly from outside. Rather it results
primarily from compensatory adjustments actively programmed within the or¬
ganism in response to the total relevant information it has available. Thus in
spite of an exquisite sensitivity to changes of many kinds, organisms maintain
their stable state with remarkable thoroughness and precision. That they sur¬
vive attests also to the reliability of their overall regulation despite failures or
irregularities in detail. . . .
While this may seem a rather loosely defined recipe for advancing science, it
does in fact provide a reasonably effective and increasingly appreciated array
of interrelated disciplines appropriate in particular ways to the formidable
problems in hand. (T. H. Waterman, 1968, pp. 3-4.)
One final fact completes this attempt to characterize the relation between
cybernetics and systems theory; that is, the interest they have engendered in
fields of study outside the natural sciences. Because these fields are not in
my area of competence, I hesitate to venture a judgment as to the influence
either cybernetics or systems theory has had. Nonetheless, there is little
question that a number of distinguished social scientists, economists, and
students of business administration have looked to these two fields for a
source of ideas and tools relevant to their particular professional needs. For
example, volume two of Progress in Cybernetics and Systems Research is
comprised of papers in Socio-Economic Systems, Cognition and Learning,
Systems Education, Organization and Management. [Trappl and Hanika,
1975.] A Survey of Cybernetics, published in 1969 and dedicated to the
CYBERNETICS 437
The reason for my defection [from the First Symposium on Cybernetics and
Systems Research] was in fact an urgent recall to Santiago from the late Presi¬
dent Salvador Allende of Chile. We had embarked six months earlier upon a
program so ambitious as to have had at least a chance of revolutionizing the
form of government on a cybernetic basis that would match the revolutionary
political intentions of that democracy. This endeavor took precedence with me
for two years, and I emerged from the experience very much changed. I
changed in my awareness of myself, of my fellow men and of political realities;
but these are not the topics that I shall discuss today. I changed also as a
technologist, in terms of confidence. For I now know that it is possible to do
what I have advocated for so many years—things which many used to say, and
some still do say, are impossible.
But the changes that bear upon the nature of this symposium have to do with
cybernetic insights themselves. There is of course no way of changing the laws
by which large systems operate; but there can be a change in one’s perception,
and a change in the depth of understanding of principles we have known about
all the time. It is of these matters that I speak today, because I know more
clearly now what I am trying to say, and because I also know more about the
direct practical relevance of these things to society at large. (Beer, 1975, p. 3.)
Those of us who have contributed to the new science of cybernetics thus stand
in a moral position which is, to say the least, not very comfortable. We have
contributed to the initiation of a new science which, as I have said, embraces
technical developments with great possibilities for good and for evil. We can
only hand it over into the world that exists about us, and this is the world of
Belsen and Hiroshima. We do not even have the choice of suppressing these
new technical developments. They belong to the age, and the most any of us
can do by suppression is to put the development of the subject into the hands of
the most irresponsible and most venal of our engineers. The best we can do is
to see that a large public understands the trend and the bearing of the present
work, and to confine our personal efforts to those fields, such as physiology
and psychology, most remote from war and exploitation. As we have seen,
there are those who hope that the good of a better understanding of man and
438 MURRAY EDEN
society which is offered by this new field of work may anticipate and outweigh
the incidental contribution we are making to the concentration of power (which
is always concentrated, by its very conditions of existence, in the hands of the
most unscrupulous). I write in 1947, and I am compelled to say that it is a very
slight hope. (Wiener, 1948 and 1961, pp. 28-29.)
CONCLUDING REMARKS
Dr. Rosenblueth has always insisted that a proper exploration of these blank
spaces on the map of science could only be made by a team of scientists, each a
specialist in his own field but each possessing a thoroughly sound and trained
acquaintance with the fields of his neighbors; all in the habit of working to¬
gether, of knowing one another’s intellectual customs, and of recognizing the
significance of a colleague’s new suggestion before it has taken on a full formal
expression. The mathematician need not have the skill to conduct a physiologi-
CYBERNETICS 439
cal experiment, but he must have the skill to understand one, to criticize one,
and to suggest one. The physiologist need not be able to prove a certain
mathematical theorem, but he must be able to grasp its physiological
significance and to tell the mathematician for what he should look. (Ibid., p. 3.)
Peter Elias
investigation devoted to it. Such in its time was the case with the theory of
integral equations, after the fundamental work of Fredholm; so it was with
information theory after the work of Shannon. [Khinchin, 1957, p. 30.]
Richard Mattessich
Additional support from the Social Sciences and Humanities Research Council of Canada is
gratefully acknowledged.
445
446
RICHARD MATTESSICH
(in the narrow sense) and systems theory becomes clear only after grasping
the reason why the former is only a part of the latter. This can best be
demonstrated by discussing the major ideas and components that modern
system theory consists of. And we may thus distinguish the following basic
ideas and notions of the systems approach:
A major concern of Aleksandr Bogdanov [1913 and 1922; 1980] and Ludwig
von Bertalanffy [1928 and 1933; 1950; 1968], the founders of system theory,
was the reductionistic and atomistic attitude of the physical sciences and the
related neopositivistic philosophy. Due to this attitude and to the fact that
the biological and social sciences were about to borrow a good deal of their
methodological inventory from the physical sciences, the phenomenon of
emergent or holistic properties, especially those of living or social systems,
was unduly neglected. But not even in physics and chemistry could those
newly emergent properties be explained in terms of mere aggregation.1 This
initiated the quest for principles of organization within the framework of
which the unique status of emergent properties could be given due recogni¬
tion and which, in time, might even acquire explanatory powers. It is for this
reason that Bogdanov’s book [1913] was called Tektologia and bears the
subtitle The Universal Science of Organization and why Bertalanffy [1968,
p. 187] speaks of an organismic revolution, the core of which is supposed to
be the notion of system. Closely connected with this is James G. Miller’s
[1978] hint that Alfred N. Whitehead, with his “philosophy of organism,’’
might be the founder or precursor of system philosophy, and Miller’s asser¬
tion that the “key concepts later accepted as basic to systems science occur
in his [Whitehead’s] writings.” (J. G. Miller, 1978, p. xiii.)2
For example, such holistic properties as the relatively harmless and useful properties of water
or cooking salt that emerged from the combination of such explosive substances as hydrogen
and oxygen or of no less dangerous substances as sodium and chloride, respectively
Personally, I am inclined to side with Wiener [1950, 1954, and 1967, pp. 27-28] in regarding
Leibniz as the true precursor of the system approach. His monadology comes fairly close to a
kind of system theory, although we believe that a monade is not to be identified with a system,
but rather with the reflective (i.e., mental or quasimental) aspect of a system.
CYBERNETICS AND SYSTEMS THEORY
3 A more correct definition is presented by Oskar Lange: “Generally speaking, ‘system is taken
to mean a set of elements together with the set of relations between the elements. The set ot
such relations (and their isomorphic transformations) is called the structure of the system.
(Lange, 1962 and 1965, p. 17.)
448
RICHARD MATTESSICH
One of the most crucial aspects found in system theory, but lacking among
the concepts of traditional science, is the notion of goal orientation or purpo¬
siveness.4 Physics and chemistry can get along without it fairly well (al¬
though the concept of entropy may weaken this statement), but there is
growing evidence that the biological sciences and certainly the social sci¬
ences need the idea of goal orientation and the related notions of preference,
value, efficiency, and so forth, most urgently. And it is no coincidence that
the biochemical theory of the Nobel laureate Manfred Eigen [1971] intro¬
duces a value function into biology that has at least some affinity with the
value function in economics.
Most of the approaches dealing with systems strongly emphasize the
notions of purpose, goal orientation, efficiency and effectiveness of goal
attainment, and so forth. But surprisingly enough (as Eden points out),
Wiener does not deal with such notions in his cybernetics, in spite of the fact
that he considered them in an earlier, coauthored paper. It is revealing for
Wiener’s thought process that the subject index of his work on cybernetics
and society [1950, 1954, and 1967] contains only a single entry under pur¬
pose, and this pertinent passage must be taken as a rejection rather than an
acknowledgment of this very notion: “Here I want to interject the semantic
point that such words as life, purpose, and soul are grossly inadequate to
precise scientific thinking.” (Wiener 1950, 1954, and 1967, p. 45; emphasis
added.) Such a radical attitude with regard to the notion of purpose is not
typical for a general-systems theoretician, mainly because the applied and
social sciences, and to some extent even the biological sciences (see Monod
[1970 and 1971] and Jacob [1971] and their project telenomique possessed by
every organism; see also Eigen [1971]) are in need of various notions of goal
and goal attainment. But some proponents of the systems approach caution
us by asserting that “biology does not have a single law statement, let alone
theory, containing the concepts of plan (or design) and purpose.” (Bunge
1979a, p. 120.) 6 ’
The fact that cybernetics (in the narrow sense) is more or less limited to this
area of control information, forms the strongest argument in favor of regard¬
ing cybernetics as a subset of systems research. The insights that (1) a
control state is dependent on the pertinent information-flow and (2) the
scientific laws governing control are universal, hence, independent of the
traditional dichotomy between living and nonliving entities, seem to form
And the hierarchy connected with this notion: state-maintaining, goal-seeking multigoal-
seeking, purposive, and purposeful. [Ackoff, 1971, p. 665.]
CYBERNETICS AND SYSTEMS THEORY 449
From the very beginning of system theory, the notion of environment has
played an essential role. Apart from the fact that it represents the logical
counterpart of the system notion, it performs important analytical and prac¬
tical services in bounding a system and determining whether certain proper¬
ties or phenomena (e.g., norms) are part of the system influencing the envi¬
ronment, or whether they impinge from outside on the system. [Mattessich,
1974a; 1978, chap. 2.] Distinctions—as those between internal versus exter¬
nal structure and relations; subsystem, system, and meta- or supersystem,
and so forth—are consequences of this basic but perhaps artificial dichotomy
between system and environment. There is little evidence that Wiener made
any conscious use of this fundamental notion. And yet, the great analytical
and pedagogic value of the system/environment dichotomy has become one
of the hallmarks of the systems approach. We encounter it everywhere,
whether we listen to a television program in which Jonathan Miller [1978]
demonstrates the analogy between the artificial inner environment of a
stratospheric airplane and the human bloodstream or whether we read a
National Aeronautics and Space Administration report on the no less
artificial environment afforded an astronaut by a space suit.
An essential feature of systems theory is, or at least ought to be, the search
for general laws underlying all systems and laws underlying more specific
but still fairly general system-types. If at least a few such nontrivial laws can
be found, the systems approach may, indeed, become a theory, although
hardly an entire scientific discipline or a superscience. But if such laws
prove to be elusive or fairly insignificant or too specific, then the systems
approach will remain a mere methodology, the ideas and tools of which will
be absorbed in time by the methodology of general science or some specific
disciplines. At this moment, it seems that the latter alternative is more likely
to occur. Although the systems approach belongs to empirical science [see
my comment on Mortazavian’s paper in this volume], the prospect of for¬
mulating fairly general and yet powerful empirical system hypotheses is not
very bright; at best, we have mathematical homologies, which, however, do
450 RICHARD MATTESSICH
not make the systems approach an analytical discipline like logic or pure
mathematics. Such homologies may best be illustrated by the isomorphism
between a mechanical and an electrical oscillator. [Compare Rapoport,
19686, p. 455.] From his introductory physics lessons, the reader may re¬
member that the force f(t) impacting a mechanical oscillator is
where L stands for inductance, R for resistance, C for capacitance (and t and
d or d~ maintaining the previous meaning). This reveals significant structural
homologies between mechanical and electromotoric force, mass and induct¬
ance, mechanical friction and electrical resistance, elasticity and electrical
capacity. Perhaps the best evidence for the nonaccidental and empirical
nature of these formal or mathematical homologies lies in the fact that the
amount of heat produced by overcoming the mechanical friction is identical
to that produced by overcoming the electrical resistance. Cybernetics and
information theory did, indeed, produce such homologies, but again, they
are fairly limited to feedback and homeostatic mechanisms, on one side, or
analogies of entropy and negentropy, on the other side. While the present
practice of system theory—especially applied system theory as practiced,
for example, by the International Institute for Applied Systems Analysis
(IIASA) in Laxenburg, Austria—operates on a much broader empirical basis
but with homologies and system models of a lower level of generalization.
CONCLUSION
These few considerations will, I hope, help to sharpen our distinction be¬
tween cybernetics and the more comprehensive systems approach. But
Eden is quite correct in pointing out that in both areas “we face a procedural
dilemma that “other domains of scholarship do not ordinarily pose.” To
regard both of them as viewpoints and methodological tools instead of scien¬
tific disciplines might be the ultimate resolution. As regards the contradic¬
tion (to use Eden’s expression) between Mesarovic—who leans toward a
purely analytical presentation of system theory (mathematical system
theory) and others, like Bertalanffy, who have a much broader vision of
CYBERNETICS AND SYSTEMS THEORY 451
5The renowned mathematician Rene Thom offers an interesting discussion about the distinction
between the concrete (morphological) and the abstract (mathematical) notion of a system and
points out that “from a mathematical viewpoint, the systemic approach leads to a rather
sketchy theory.” (Thom, 1980, p. 3.) , ,
6This point of view comes close to that of Rapoport, who suggests that “general systems theory
is best described not as a theory in the sense that this word is used in science but, rather as a
program or a direction in the contemporary philosophy of science. " (Rapoport, 1968b, p. 4,2.)
7There is a crucial difference between this system definition and that of Mortazavian, admit¬
tedly based on that of Mesarovic [1964a], Apart from the fact that we distinguish between a
concrete and an abstract system, our definition of the latter is first of all a special kind of set of
objects and not merely a set of relations between those objects. As can already be seen from
footnote 3,1 respect the importance of this set of relations, but it represents the structure of the
system, not the system itself.
452 RICHARD MATTESSICH
KIn connection with Eden’s paper and my present comment, it is interesting to note that the
British Journal Kybernetes (Thales Publications Ltd.) initiated the (1977) Norbert Wiener Com¬
petition under the closely related title. “The Unity between Cybernetics and General Systems
Theory," published in its vol. 8 [1979] with pertinent contributions by Counelis (third prize),
Gergely, Majumder (first prize), Mayne, Stanley-Jones, and Zeleny (second prize). I am grateful
to George Gorelik, University of British Columbia, for drawing my attention to this particular
publication.
CYBERNETICS IN THE INFORMATION
DISCIPLINES
Manfred Kochen
Murray Eden concludes his essay on the history of cybernetics with the
implication that the only function of cybernetics was to diffuse the principles
of communication and control engineering into the culture of science and
technology and to integrate itself into it. That has in fact been its primary
effect. But it had more than one intended function. Another was to identify
analogous or identical concepts in diverse fields and analyze those that are
essentially mathematical by appropriate mathematics. It has not yet per¬
formed this function with equal success. The prospect of that happening in
the next decade or two is the first point on which I would like to comment.
The second point concerns Dr. Rosenblueth’s insistence on the need for
teams of scientists to establish fruitful connections among principles of com¬
munication and control engineering and living systems, as reiterated by
Wiener in Cybernetics and quoted by Eden. The need for teamwork is more
urgent today than ever before, not only for interdisciplinary scientific re¬
search but for many survival-related activities. Despite the trend toward
segregation into increasingly narrow specialties, as indicated by Eden, there
are now techniques and signs of hope that this trend could be reversed.
My third point is stimulated by the phrase “the map of science” in the
same quotation of Rosenblueth by Wiener. Cybernetics did not become a
distinctive discipline or specialty as pointed out by Eden, though he hints
that like other intersections of different disciplines, in time it might become
yet another specialty. Under what conditions do new scientific specialties
come into being, survive, wax or wane, merge with others, and die out?
What is the likely role of cybernetics in the emerging information disci¬
plines? A catalyst? One of them? A substrate?
It is quite possible that a new discipline of scientific planning to steer and
manage complex human systems will emerge to perform one of the major
intended functions of cybernetics: interdisciplinary teamwork in both scien¬
tific research and social development on a set of concepts related to syn¬
ergetics. [Haken, 1980, p. 121.]
453
454 MANFRED KOCHEN
1 Marvin Minsky and Seymour Papert, Perceptrons [1969]. The invention of the perceptron was
reported by Frank Rosenblatt [1958] and developed more fully in Rosenblatt, Principles of
Neurodynamics [1962],
CYBERNETICS IN THE INFORMATION DISCIPLINES 455
engineers have been doing in designing reliable computer systems. John von
Neumann was the first to see the essentially mathematical problem common
to both engineered and living systems, and this has led to a line of research,
the full potential of which has yet to be developed, [von Neumann, 1956;
Moore and Shannon, 1956; Kochen, 1959; S. Winograd and Cowan, 1963.]
Though the stimulus to several other promising areas of new mathematics
could be traced to Wiener and cybernetics, only two will be mentioned. The
first is the use of probabilistic and computational methods in even as pure a
branch of mathematics as number theory.
To test whether a number n, less than, say, 2400 —593, is prime, select a
random sample of 50 numbers a\, , a50, each less than n, and test if a, is a
witness to the nonprimality of n by checking that n does not divide a1~ 1 for
each i. If all a, fail this test, call n prime; the probability of error is less than
(1/4)50 or (1/2)100. The computational complexity of such an algorithm is far
less than that of a nonprobabilistic one. Somewhat similarly, the four-color
conjecture has been settled with the help of a computer analysis of cases.
While such methods are not universally accepted in the mathematical com¬
munity, they are increasingly used by its leaders and may contribute to the
way we think in mathematics. The idea of thinking stochastically is as pro¬
found as the Newtonian idea of thinking continuously or the Poincare idea of
thinking qualitatively. One of the roots of thinking stochastically is in a
central tenet of Wiener’s formulation of cybernetics: Our capacity to control
is inherently incomplete and necessitates a statistical approach.
The same tenet may also have been a precursor to “fuzzy-set theory.”
Control engineers, interested in designing a system that could park a car, for
example, did not believe that the classical mathematical concepts used in
designing control systems were appropriate. [Zadeh, 1965; Lowen, 1981;
Goguen, 1967; Smets, 1982.] A person performs the task in what seems to be
a totally different way. He or she does not measure the variables fed back to
the visual system with any greater precision than he or she specifies control
variables with the motor system. Nor did probabilistic concepts appear to be
fruitful. Instead, a new kind of mathematics was spawned that started from a
generalization of the characteristic function of a set, replacing its two-valued
range by the interval [0, 1]. Thus, the characteristic function of the set of
large numbers for a particular judge was interpreted as the degree of mem¬
bership in the class (fuzzy set) of large numbers assigned to any number x by
that judge. For x - 1 to 10, it might be 0, for x = 11 to 50, 0.1; for x = 51 to
100, 0.2, and so forth; and for x > 1000, it might be 1. Fuzzy-set theory has
stimulated contributions of a purely mathematical nature, though its future
standing among the mathematical specialties cannot yet be forecast.
It would appear that contemporary young scientists are more broadly and
deeply trained than their predecessors. Nearly every good science student
456 MANFRED KOCHEN
'
CYBERNETICS
The View from Brain Theory
Michael A. Arbib
Murray Eden starts his review with the sentence: “Norbert Wiener, in his
book entitled Cybernetics, or Control and Communication in the Animal
and the Machine, did not define explicitly the word he believed he had
coined.” Surely, the book’s title is, at least implicitly, a definition. How¬
ever, Eden is correct to stress that Wiener’s book is intensely personal and
brings together a vast array of hitherto disparate topics. He places Wiener’s
book in a historical perspective that goes back to Hero of Alexandria and
reaches the present day via Descartes and Ampere, among others. I trace a
similar history in my companion article (“Cognitive Science: The View from
Brain Theory”), where I argue that the history of cybernetics is, in large
part, also the history of cognitive science. This is consistent with Eden’s
discussion of the permeation of psychology by cybernetic concepts and
counterbalances Pylyshyn’s overemphasis on mathematical logic and
artificial intelligence (AI) as driving forces in the development of cognitive
science.
In our department at the University of Massachusetts at Amherst, we use
the word cybernetics to refer to a conjoined study of brain theory and AI.
[Lesser, 1981-1982.] However, there are many experts in AI who regard
cybernetics as an outdated, outmoded term. They insist that there is no need
for an umbrella term that relates AI to brain theory. Eden notes that others
use the word cybernetics in the sense of any study of complex systems that
is aided by a computer—a broad usage encompassing even mathematical
economics. (This is particularly so in the Soviet Union, see [Arbib, 1966].)
Finally, he suggests that, while some workers still use the word cybernetics,
This paper refers to research supported in part by NIH grant NS 14971 and NSF grant MCS
8005112 at the University of Massachusetts. Portions of the paper are adapted from a paper
prepared for a symposium marking the 25th anniversary of the publication of Wiener’s Cyber¬
netics. [Arbib, 1975b.]
459
460 MICHAEL A. ARBIB
much of the original energy of the field has moved to a vast range of descen¬
dant disciplines. I think that this is true but still believe the umbrella term
serves well to remind workers in different disciplines of the range of com¬
mon concepts they share.
Statistical Mechanics
As Eden stresses, Wiener’s book reflects strongly the influence of his rich
experience in using statistical methods in control, communication, and fil¬
tering problems. In World War II, Wiener had studied problems of tracking,
interpolation, smoothing, and prediction, particularly in relation to anti¬
aircraft gunnery. The great insight of this study had been that we should not
ask, “Here is a trajectory; now what is the best way of continuing it?’’
without having some information about what ensemble that trajectory be¬
longs to. Having characterized an ensemble by the noise statistics of a sto¬
chastic process, we could make meaningful predictions, such as, “This is
the proposed continuation of the trajectory to which the possible continua¬
tions will, in some well-defined statistical sense, come as close as possible.’’
In his book, Wiener sought to unify these mathematical studies with his
discussions with psychologists and physiologists, to lay the foundations for a
new field, to be called cybernetics. Let me quote a passage from Wiener’s
book where he presents his views:
Wiener knew from his engineering experience that if the gain is too high in a
feedback system, it will go into unstable oscillations. This led him to suspect
that there might be a form of human brain damage that would cause the
human’s limb to go into violent oscillations when the patient tried to pick
things up. Sure enough, that turned out to be the case—his clinician friends
showed him such an effect in patients with cerebellar lesions. This is perhaps
the root of the whole development of control-theoretic analysis of complex
biological systems. [Milsum, 1966.]
Wiener’s discussion of filtering, feedback, and stability does not use the
idea of a state variable; rather, he discussed how, given the infinite history of
the machine’s input, we could specify the infinite story of its output. He did
not ask such questions as “What is going on inside the box?” or “How can
we relate the internal dynamics to the input/output behavior?” It is Rudolf
Kalman, perhaps as much as anyone, who has made state-variable thinking
mandatory in system theory. [Kalman, Falb, and Arbib, 1969.]
As Eden notes, Wiener suggested that “the computing machine must
represent almost an ideal model of the problems arising in the nervous
system.” This might seem at cross-purposes with the previous emphasis on
statistical mechanics. However, Kalman and Richard Bucy provide a rap¬
prochement, for they showed how to replace the history-based statistical
extrapolation method, known as the Wiener-Hopf filter, by the state-based
recursive estimation technique, known as the Kalman-Bucy filter—which,
implemented in computer hardware, is an integral part of guidance systems
462 MICHAEL A. ARBIB
in today’s planes, missiles, and spacecraft. [Kalman and Bucy, 1961.] Inter¬
estingly, the extrapolation problem has recently resurfaced in brain theory,
with the look-ahead modules of Andras Pellionisz and Rodolfo Llinas,
though, unfortunately, they have not tried to apply these earlier successful
techniques from system theory. [Pellionisz and Llinas, 1979.]
Internal Models
When a system is carrying out some control task, its parameters may not be
quite matched to its job. This suggests the need for what Wiener calls infor¬
mative feedback and what most of us today would call either model-
referencing or feed-forward. Here, we augment the core feedback-system by
using an identification procedure to upgrade the parameters of the core
system at the same time that it is locked into a feedback loop for ongoing
control.
What is very interesting from our present perspective is that despite this
feed-forward idea, there is no mention in Wiener’s book of the work of the
psychologist Kenneth Craik. Craik wrote perhaps the first work in cyber¬
netic psychology—though, of course, not yet with that name. [Craik, 1943.]
He looked at the nature of explanation and stressed the idea of our building
in our heads a complex model of our world, not in the cardboard-and-glue
sense, but, nonetheless, a model that could modulate our input/output be¬
havior to match our experience of our world. This idea, not in n-dimensional
differential-equation format, but in a very general data-structure format, has
become crucial to much of the exciting work in artificial intelligence. We saw
Shakey the robot using visual input to build an internal model of the world
about which it will navigate [Nilsson and Raphael, 1967], while Terry
Winograd based his approach to linguistics on having the computer able to
refer to a model of its simple world of blocks on a table top when “disambig¬
uating” sentences that it is using. [Winograd, 1972a.] The last ten years of
AI have been dominated by a retreat from general methods and an increasing
emphasis on techniques for representing specialized domains of knowl¬
edge—the so-called “expert systems.” [R. Davis and Lenat, 1982; Short -
liffe, 1976.]
So it goes on. We could relate the work in neurophysiology since 1948 to
Wiener’s interests; we could follow Wiener’s notions on intermodality cod¬
ing; and we could discuss at length the social problems that have become an
inescapable reality since 1948. Nevertheless, here I must leave this retro¬
spective look at Wiener’s book. What I hope to have done is confirm Eden’s
demonstration that Wiener had a broad intellectual range and many of the
questions Wiener raised are still very much alive, though transformed and
enriched by the developments—both technical and conceptual—of the last
35 years.
CYBERNETICS: THE VIEW FROM BRAIN THEORY 463
BRAIN THEORY
A few years ago, I discussed two specific aspects of my research that may be
viewed as having descended from Wiener’s work (among others). [Arbib,
1975b.] First, I explored some contributions in modern system theory—the
descendant of Wiener’s control theory—by giving a nonmathematical pros¬
pectus for the highly mathematical study of “machines in a category” that
Ernest Manes and I conducted at the University of Massachusetts (reviewed
by [Arbib and Manes, 1980]) and, second, I turned to a discussion of brain
theory. Rather than repeat here the discussion of machines in a category, I
want simply to recall its conclusion. We concluded that our general insights
into the functioning of systems had come about only because we could use
category theory to build on the theories of sequential machines, linear sys¬
tems, and tree automata. We had a sufficiently rich set of particular theories
as examples, and we had a sufficiently general mathematical theory to pro¬
vide the language for their unification. Brain theory, unlike mathematical
systems theory, is only now building up a reasonable stock of good examples
of theories of limited applicability, which may serve as a basis for developing
powerful general techniques that can be brought to bear on a wide variety of
specific problems relating the physiology and anatomy of neural circuitry.
For example, early studies of the role of lateral inhibition in, for example,
contrast enhancement in sensory systems can now be related to models of
mode selection, where emphasis is on directing attention to one mode, fea¬
ture, or set of features of activity, rather than on preprocessing an array for
further computation. [Ratliff, 1965.] These models are in turn part of a broad
class of models dealing with competition and cooperation in neural nets.
One of the first models of decision-making in neural circuitry to explicitly
opt for cooperative computation was the S-RETIC model. [Kilmer, McCul¬
loch, and Blum, 1969.] It suggested how the reticular formation of the brain
stem might switch the organism’s gross state from one overall mode of
behavior in the organism to another. Based on the anatomical work of
Madge and Arnold Scheibel, Kilmer and McCulloch believed that a rea¬
sonable structural simplification of the system was a stack of poker chips.
[Scheibel and Scheibel, 1958.] (The Scheibels’ work may be cited as one of
the earliest contributions to modules of neural structure intermediate be¬
tween brain region and neuron. [Mountcastle, 1978.]) In the S-RETIC
model, each module in the stack receives a sample of the overall system
input and on the basis of that sampling, assigns weights that provide likeli¬
hood estimates for each of the different modes. Initially, these estimates are
based only on a local sample. However, the modules are coupled in such a
way that each module readjusts its weights on the basis of activity in other
modules to which it is connected. Kilmer and McCulloch suggested a con¬
nection scheme that would lead to eventual consensus, with a majority of the
modules assigning the greatest weight to a single mode—all without any
executive control.
464 MICHAEL A. ARBIB
We now turn to the topic of segmentation of visual input from two eyes on
the basis of depth cues. Bela Julesz designed random-dot stereograms in
which each eye receives a totally random pattern but in which different
regions in the two inputs are identical save for a shift in position, yielding a
different disparity on the two retinas. [Julesz, 1971.] Although such a pattern
for a naive subject can initially appear to be visual noise, disparity-matching
takes place eventually, and the subject perceives surfaces at different
depths. Neurophysiologists have found that cells in a cat’s visual cortex are
tuned for retinal disparity, and similar cells are posited in the human. What
presumably causes the initial perception of visual noise is that in addition to
the correct correlation of points in the two retinas, there are many spurious
correlations, and computation is required to reduce them.
Parvati Dev was one of several workers to propose that the cells of a
given disparity be imagined as forming a population arrayed in a spatial map
corresponding to the map of visual direction. [Dev, 1975.] Connections be¬
tween cells could then be arranged so that nearby cells of a given disparity
would be mutually excitatory, whereas cells nearby in visual direction but
different in disparity would have inhibitory interaction. In this way, the
activity of the array would be organized into a pattern where cells of only
one disparity-type would be highly active in each region of visual direction.
As a result, the visual input would eventually be segmented into a number of
distinct surfaces. (See David Marr’s posthumously published magnum opus.
Vision, for an influential approach to such problems. [Marr, 1982.])
In the stereopsis model, then, we have competition in the disparity dimen¬
sion and cooperation in the other dimensions. We note the similarity to S-
RETIC, where the cooperation dimension is the row of modules, and com¬
petition is between modes rather than disparities. Other workers have
shown that a number of interesting features of visual psychophysics can be
captured in a fairly simple neural network in which we have two interacting
populations of cells—excitatory and inhibitory—with plausible interconnec¬
tions. Such a model exhibits hysteresis, one of the most famous psychophys¬
ical manifestations of which is seen in fusion. Such hysteresis phenomena
are also central to the study of stereopsis. Informal observations of this kind
have laid the basis for rigorous mathematical analysis, which also subsumes
an early model of frog prey-selection. (We note that these neural mecha¬
nisms are similar to the relaxation methods for segmentation and region
labeling in machine vision. [Hanson and Riseman, 1978.]) A recent United
States-Japan seminar brought together applied mathematicians, brain theo¬
rists, and neuroscientists to present a body of experimental and theoretical
techniques in brain theory. The papers, collected by Amari and Arbib, give a
good picture of both successes in building general theory and also the many
strands of theory and experiment that still need to be tied together—or
trimmed away completely. [Amari and Arbib, 1982.]
Another area of brain theory of continuing interest is the control of move¬
ment, including Nikolai Bernstein’s theory of synergies (motor schemas)
CYBERNETICS: THE VIEW FROM BRAIN THEORY 465
Murray Eden
The three of us felt that this new point of view merited a paper, which we wrote
up and published. (Wiener, 1948, p. 8.)
I mention this matter because of the considerable, and I think false, hopes
which some of my friends have built for the social efficacy of whatever new
1 He is in good company. John von Neumann began his seminal paper on probabilistic logics
with the statement: “Our present treatment of error is unsatisfactory and ad hoc. It is the
author’s conviction, voiced over many years, that error should be treated by thermodynamical
methods, and be the subject of a thermodynamical theory, as information has been, by the work
of L. Szilard and C. E. Shannon. The present treatment falls far short of achieving this, but it
assembles, it is hoped, some of the building materials, which will have to enter into the final
structure.” (von Neumann, 1956, p. 43.)
470 MURRAY EDEN
ways of thinking this book may contain. They are certain that our control over
our material environment has far outgrown our control over our social environ¬
ment and our understanding thereof. Therefore, they consider that the main
task of the immediate future is to extend to the fields of anthropology, of
sociology, of economics, the methods of the natural sciences, in the hope of
achieving a like measure of success in the social fields. From believing this
necessary, they come to believe it possible. In this, I maintain, they show an
excessive optimism, and a misunderstanding of the nature of all scientific
achievement. (Ibid., p. 162.)
We note that in the passage quoted, Wiener implied that his were “the
methods of the natural sciences.”
Peter Elias has contributed a valuable addition to the brief history I have
presented. His comments on the Soviet response to the idea of cybernetics
are particularly welcome. His explanation of the difference between Soviet
and Western usage of the term cybernetics is both simple and plausible.
Elias points out that there were, almost from the beginning, different at¬
titudes toward cybernetics in the United States and Europe. Sensitivities
with respect to terminology were heightened even within the community of
European workers. Elias cites W. Ross Ashby as the first author of a text on
cybernetics. [Ashby, 1956.] Curiously enough, Ashby, in his earlier Design
for a Brain makes only one reference to cybernetics. [Ashby, 1952 and
I960.] Colin Cherry, another early English worker in the field, much pre¬
ferred the label communication theory. For him, a part of the theory was
“the theory of feedback (sometimes called cybernetics).” (Cherry, 1957, p.
21.) Later he writes:
This wider field [communication theory and physics] ... is referred to, at least
in Britain, as information theory, a term which is unfortunately used elsewhere
synonymously with communication theory. Again, the French sometimes refer
to communication theory as cybernetics. It is all very confusing! (Ibid., p. 216.)
Max Bense has contrasted the classical (Archimidean) world emphasizing en¬
ergy and work with the nonclassical (Pascalian) world emphasizing information
and communication. Although some aspects of the nonclassical world have
been known for a long time, it has become accessible to systematic research,
transcending all narrow overspecialization, only thanks to two essentially
mathematical theories, information theory on the one hand and theoretical
cybernetics on the other. ... It is the task of information theory to open up to
quantitative and structural comprehension either communication between per¬
son and person, as it manifests itself in exchanges of symbols, or communica¬
tion between man and the world, which essentially amounts to observation; by
CYBERNETICS: CLOSING THE LOOP 471
The field of scholarship that the commentators and I are trying to charac¬
terize—I dare not try to assign it a name—appears to encourage terminolog¬
ical invention and yet greater confusion. For example, Kochen, disagreeing
with Minsky as quoted by me, notes that perceptron-related work is being
actively pursued. The work Kochen cites is quite close to work that Arbib
includes within artificial intelligence. Minsky’s definition is implied by the
title of his book (with Papert), Perceptrons—An Introduction to Computa¬
tional Geometry. [Minsky and Papert, 1969.] On the other hand, Rosenblatt,
who coined the term perceptron, identifies it with a class of “brain models.”
[Rosenblatt, 1962.]
Mattessich has got it right when he attributes this tangle to “the human
urge to devise impressive labels.” I share with all four commentators a
measure of regret that the taxonomy of the information sciences is still so
ambiguous. Nevertheless, insofar as progress in scientific understanding is
concerned, there is no great loss. Fortunately, whatever the global label,
each particular research report is ultimately verified by the tests of the
underlying domain of scholarship. To put it somewhat simplistically, if the
work is mathematical, we check the theorems; if it is a model, we check to
see whether it fits empirical observations. Perhaps we should be guided by
Shakespeare: “What’s in a name? That which we call a rose by any other
name would smell as sweet.”
■
*•
SECTION 7
THIRTY YEARS OF
INFORMATION THEORY
Myron Tribus
This paper is concerned with the impact of Shannon’s 1948 paper “A Mathe¬
matical Theory of Communication,” which was focused on the field of com¬
munications. [Shannon, 1948.] I especially wish to comment on the influence
of Shannon’s paper in fields other than communications. I approach the task
with some hesitation: In 1961 Shannon, in a private conversation, made it
quite clear to me that he considered applications of his work to problems
outside communication theory to be suspect and he did not attach funda¬
mental significance to them.
Despite Shannon’s misgivings, as time goes on, his great contribution to
the literature has made its existence felt in an ever-increasing number of
fields. Shannon’s clarification of the concepts of uncertainty and information
and the ability to give a quantitative measure to these concepts has served as
a powerful stimulant on the imagination of others. Of course, there have
been numerous unjustified excursions in the name of information theory, but
my studies show many solid accomplishments clearly inspired by this fa¬
mous paper.
To test the spread of Shannon’s influence, I thumbed through the En¬
gineering Index and found that the first entry under the heading Information
Theory occurred in 1951, only three years after his paper appeared. In 1952,
there were eight entries; in 1953, there were thirteen entries, including three
symposia and a reference to a bibliography containing over 1000 references
to information theory. By 1958, ten years after Shannon’s paper, the En¬
gineering Index contained several pages of titles covering papers in the field
of information. Many of the entries were references to symposia that them¬
selves contained between fifteen and twenty papers on information theory. I
can report also that the Engineering Index listing under the title Information
Theory is incomplete. For example, in 1961, I prepared a paper on informa-
This is a slightly modified version of a paper first published as an article in Levine, Raphael D.,
and Tribus, Myron, eds., The Maximum Entropy Formalism (Cambridge, Mass.: MIT Press,
1979).
475
476 MYRON TRIBUS
tion theory and thermodynamics. It was listed by the Index under Thermody¬
namics but not under Information Theory. There is, therefore, no easy way
to track down all of Shannon’s influence. For example, under the heading
Entropy, the Engineering Index for 1961 merely says “See Thermodynam¬
ics." Once we know the field that was influenced, however, it is fairly easy
to document the influence.
It is rather difficult to say where the field of information theory ends and
other fields begin. Certainly, information theory overlaps significantly with
many other fields. Define S — entropy, K = constant, pt = probability of the
truth of assertion i. For the purposes of this paper, I shall take the view that
any field of inquiry that uses the function
(1)
(or its continuous analog) in one of the following three ways is using informa¬
tion theory: (1) as a criterion for the choice of probability distributions; (2) to
determine the degree of uncertainty about a proposition; and (3) as a mea¬
sure of the rate of information acquisition.
Claude Shannon was not the first to use the function - X,-p,- In ph for this
function (or its continuous analog) had been used as long ago as 1872 by
Boltzmann. [Boltzmann, 1872.] What Shannon did was to give a universal
meaning to the function - In p; and thereby make it possible to find other
applications. Warren Weaver, in his popularization of Shannon’s work,
foresaw the widespread influence it was bound to have. Weaver understood
that any paper that clarified our understanding of knowledge was certain to
affect all fields that deal with knowledge.
In the 1961 interview with Shannon to which I referred, I obtained an
anecdote worth recording here. I had asked Shannon what his personal
reaction had been when he realized he had identified a measure of uncer¬
tainty. Shannon said that he had been puzzled and wondered what to call his
function. Information seemed to him to be a good candidate as a name, but it
was already badly overworked. Shannon said he sought the advice of John
von Neumann, whose response was direct, “You should call it ‘entropy’ and
for two reasons: First, the function is already in use in thermodynamics
under that name; second, and more importantly, most people don’t know
what entropy really is, and if you use the word ‘entropy’ in an argument you
will win every time!’’
THERMODYNAMICS
from the Greek language to mean transformation and defined it via the
equation
dS = -42-, (2)
dS = - AT ^ In Pi dp^ (4)
i
dS = -42L. (5)
In 1953, Brillouin formally stated the case for a close connection between
these two entropies. [Brillouin, 1953.] He followed his early paper with
several others and with his famous book on Science and Information
Theory. [Brillouin, 1962.] In these publications, Brillouin showed there was
an intimate connection between entropy and information, but he did not
think it necessary to define one in terms of the other. Brillouin treated the
field of thermodynamics as correct and preexisting, and he treated Shan¬
non’s information theory as a correct theory. He then proceeded to demon¬
strate a consistency between the two entropies.
As early as 1911, Van der Waals had proposed that there ought to be a
connection between Bayes’s equation in probability theory and the second
law of thermodynamics. [Van der Waals, 1911.] As we now know, Shan¬
non’s measure can be derived through the use of Bayes’s equation. [Tribus,
Evans, and Crellin, 1964; Good, 1950.] In 1930, G. N. Lewis, in a discussion
of irreversibility, had written “Gain in entropy means loss of information—
nothing more.” [Lewis, 1930.] As I mentioned earlier, Boltzmann had used
the //-function as early as 1872.
In 1938, Slater based his book on the definition
(6)
choosing this expression for entropy over the more conventional one,
S = k\n W, (7)
STATISTICS
(8)
r = 1, 2, ... m (9)
(10)
(12)
r= 1
480 MYRON TRIBUS
The Lagrange multipliers are related to the given data by the equation
dXq
<Sr> • (14)
dXr
d2X0 d2X0
var (gr) covar (grgs). (15)
dX dXrdXs
For other relations, see Jaynes or Tribus. [Jaynes, 19576; Tribus, 1969.]
Jaynes’s principle shows that if Shannon’s measure is taken to be the mea¬
sure of uncertainty, not just a measure, the formal results of statistical-
mechanical reasoning can be carried over to other fields.
Jaynes proposed, therefore, that in problems of statistical inference
(which means problems for which the given data are inadequate for a deter¬
ministic prediction of what will [or did] happen), the probabilities should be
assigned so as to maximize
subject to
X P> = 1 0?)
i
and any other given information. Jaynes thus essentially defines probabili¬
ties via Shannon’s measure. From this basis, we can see that any field that
uses probability theory is a candidate to be influenced by Shannon’s work.
I must confess that in 1957 I felt there was no explanation to be had and
complained publicly that Shannon had confused things by calling two differ¬
ent ideas by the same name! I was forced to retract that ill-considered
judgment in a footnote to my paper “Thermodynamics: A Survey of the
Field” and wish now that I had retracted the whole paper! [Tribus, 1958.]
THERMODYNAMICS (AGAIN)
An explanation of the two entropies can be had only if we clarify our ideas
about what we mean by (1) heat dQ, (2) temperature T, and (3) probability p,.
What Shannon’s work did, as interpreted by Jaynes, was to make it
possible to take the information-based entropy as the primitive concept and
explain dQ, T, and p, in terms of S.
Jaynes’s initial publication demonstrated that this approach permitted a
derivation of statistical mechanics. He later went on to give illustrations of
THIRTY YEARS OF INFORMATION THEORY 481
RELIABILITY
STATISTICS (AGAIN)
which simply tells us to compare the sum of entropies in the margins of the
contingency table with the entropy of the center and see where the informa¬
tion resides. It is especially satisfying to be able to report that the first-order
approximation to the right-hand side of Equation (18) is the familiar chi-
square statistic. [Tribus, 1969, chap. 6.]
PSYCHOLOGY
NEUROLOGY
Measuring the ability of people and animals to transmit and receive informa¬
tion through various sensory channels is an obvious way of applying infor-
THIRTY YEARS OF INFORMATION THEORY 483
mation theory, and Shannon’s work could be taken over almost unchanged
for this purpose. George Miller has pointed out, however, that the concep¬
tual foundations of information theory have as much importance as the
numerical measure. They provide a method for organization and patterning.
Miller points out that concepts of gestalt psychology may be put to test via
information-theoretic concepts. [Miller, 1953.] The view that psychoneural
activity is the economical encoding of experience may also be attributed
directly to Shannon’s influence.
It may properly be said that the penetration of field X by field Y is com¬
plete when the concepts of field Y are used to prepare a textbook for stu¬
dents in field X. In this sense, the appearance of Wendell Garner’s book in
1962, explaining psychological concepts via information theory, was a mile¬
stone. [Garner, 1962.]
SOCIAL SCIENCE
OTHER FIELDS
rial populations in the Beaufort sea [Kaneko, Atlas, and Krichevsky, 1977],
information-processing of schizophrenics [Wijesinghe, 1977; Neufeld, 1977],
molecular biology [Berger, 1977], diagnostic value of clinical tests, and the
nature of living systems [J. G. Miller, 1976].
The ideas have also spread over the globe. J. O. Sonuga in Lagos,
Nigeria, uses the entropy principle to analyze runoff. [Sonuga, 1976.] In the
Netherlands, Van Marlen and Dijkstra use information theory to select
peaks for retrieving mass spectra. [Van Marlen and Dijkstra, 1976.] At the
Free University of Berlin, Eckhorn and Popel apply information theory to
the visual system of a cat. [Eckhorn and Popel, 1974.] In Canada, Reilly and
Blau use entropy in building mathematical models of chemical reacting sys¬
tems [Reilly and Blau, 1974], as does Levine in Israel [Procaccia, Shimoni,
and Levine, 1976]. And I suppose we can say the subject has really arrived
when it is referenced by the U.S. Congress Office of Technology Assess¬
ment! [Coates, 1976.]
The future we can perceive but dimly. But surely Shannon’s paper de¬
serves to be read by future generations with the same sense of intellectual
adventure that we associate with the great works of the past.
THE WIDER SCOPE OF
INFORMATION THEORY
Donald M. MacKay
Tributes to the genius of Claude Shannon are well merited, and as one of his
earliest admirers, I gladly echo those paid by Myron Tribus. My purpose
now, however, is to say something about those aspects of the concept of
information that Shannon’s theory (for good and clearly stated reasons)
chose to ignore. For it is, indeed, the case, as Shannon so often insisted, that
his theory did not define the concept of information at all. [Shannon, 1951, p.
219.]
What Shannon’s famous £p,- log (1/p,) defines is the mean statistical unex¬
pectedness of an item of information selected from a given ensemble; and
this measure is benignly agnostic as to the meaning of the items of informa¬
tion selected in a communicative process except insofar as it affects their
prior probabilities of selection.
For this reason, I am not so sanguine as Tribus in his paper that “Measur¬
ing the ability of people and animals to transmit and receive information
through various sensory channels is an obvious way of applying information
theory.” The basic problem, of course, is that the probabilities p,- in Shan¬
non’s formula presuppose a well-defined ensemble; but in the case of a
human subject, it is not at all obvious how we should define the operative
ensemble, let alone its relative frequencies. It is still less clear that what we
really want to understand about human information-processing is always
illuminated by appeal to the particular measure of unexpectedness that
Shannon developed for the assessment of communicative channel capacity.
To try to translate every reference to information (whether in biology or
elsewhere) into a statement about unexpectedness would be as inept, and as
conceptually Procrustean, as translating all references to a house into state¬
ments about its size.
485
486 DONALD M. MACKAY
Information theory, in the more general sense it has developed over the past
forty years, is concerned with all processes in which the spatio-temporal
form of one set of objects or events (at A) determines the form of another set
(at B) without explicit regard for the energetics involved. These are situa¬
tions in which we say that information flows from A to B. In this operational
context, then, we can define information as that which determines form, in
much the same way as force is defined in physics as that which produces
acceleration.
In all cases, information is operationally defined (like energy) by what it
does. [MacKay, 19546; 1969.] But whereas the work done by energy is
physical in character, the work done by information is logical work. In
talking about information, there is always a suppressed reference to a third
party, since, as in the physical theory of relativity, we have to relate our
definitions to an observer, actual or potential, before they become opera¬
tionally precise.
In communication engineering, the form of a received message can be
determined in one of two ways: (1) by a process of construction, as when the
form of a television picture is built out of light-spots; or (2) by a process of
selection from a range of preconstructed forms in obedience to a code signal,
which has no necessary isomorphism with the form selected but merely
specifies its address, as in Morse code.
Shannon’s theory was concerned with processes of type (2), where the
prime question is how many (or how few) elementary code signals can be
sufficient (on average) to identify each form required to be selected. His
answer—brilliantly conceived and elaborated—was that the average num¬
ber of elementary signals required is proportional to the weighted mean
unexpectedness log (1 //?,); but that in “noisy” situations, reliability can
be increased by using more signals per selection and designing a redundant
code to allow errors to be corrected automatically. One of Shannon’s great¬
est achievements was to show in a precise sense that a given noise level
could, in principle, be offset by introducing a corresponding degree of redun¬
dancy, so as to achieve practically error-free communication of sufficiently
long messages.
An ingenious analogue of this theorem, due to Winograd and Cowan, has
shown how structural redundancy in a computing network can be exploited
to achieve error-free computation with unreliable elements. [S. Winograd
and Cowan, 1963.] This has important implications for the theory of informa¬
tion-handling networks, including the central nervous system, where distrib¬
uted processing may be organized so as to optimize reliability.
Processes of type (1), in which forms are not selected but constructed, are
typical of scientific experimentation. For example, data are accumulated and
processed into the form of a graph, or light is transmitted through a micro¬
scope or telescope and forms an image on a photographic plate. In this
THE WIDER SCOPE OF INFORMATION THEORY 487
A major topic of discussion at the 1950 meeting was the relation between
Shannon’s measure and the thermodynamic concept of entropy, first in¬
troduced into the theory of heat engines in the nineteenth century as a
measure of the unavailability of thermal energy. The expression, derived
from the statistical theory of gas molecules as a measure of their nearness to
thermal equilibrium (and, hence, of their inability to work a heat engine),
contains 1 pj log (1/p,) as a factor. What it signifies here, as Tribus has
explained, is the amount of information (in the Shannon-Wiener sense of
selective information content), concerning the detailed microstate of the
molecules, that remains undeducible from the macrodescription available (in
terms of pressure, volume, temperature, etc.)—in short, the extent to which
the microstate is underspecified.
Now, every unit of missing information (in this sense) means that a corre¬
sponding amount of energy cannot be pinned down to do work. It turns out
THE WIDER SCOPE OF INFORMATION THEORY 489
to these questions, rather than from our armchair theorizing, that our knowl¬
edge grows. Equally clearly, it is the quality of our thinking in the armchair
that will determine whether the questions we ask receive a worthwhile
answer.
CONCLUSION
tion per se from various measures of its amount. It should be added that
some of the most generally useful qualitative concepts of information en¬
gineering cited by Tribus, such as feedback and feedforward, have quite
independent origins and owe nothing to the mathematical theory of informa¬
tion-measurement. [MacKay, 1954 b.]
This said, however, I agree with Tribus in expecting information-
theoretical concepts, both qualitative and quantitative, to play an increas¬
ingly fruitful role in the construction, testing, and (doubtless) demolition of
theories of brain function and kindred processes in a multitude of other
contexts.
INFORMATION THEORY
IN PSYCHOLOGY
George A. Miller
Some enterprises are defeated by success. I like to think that the application
of information theory to experimental psychology illustrates this oxymoron.
It is true that Shannon’s measure of information—the mean value of the
logarithmic probability [see Tribus’s Equation (1)]—is seldom seen these
days in psychological publications. It enjoyed a vogue in the 1950s, became
widely known among experimental psychologists in the early 1960s, and
disappeared by the 1970s. Students, noting this history, usually assume that
it disappeared either because it was proven false or, worse, because the
intellectual position it represents became uninhabited. Neither assumption is
correct.
What students cannot appreciate is what life was like before information
theory. Perhaps an example will help. I recall a series of experiments in¬
tended to determine the absolute threshold for the perception of visual form.
Alternative shapes were displayed in near darkness, and illumination was
gradually increased until viewers were able to identify the shape correctly.
Different experimenters obtained wildly different results. Today, it is almost
inconceivable that such experiments could have been run without control¬
ling the viewer’s expectations of what shapes might occur and without in¬
structing the viewer on the acceptable set of alternative responses. No won¬
der the results were all over the lot!
The point is that before anyone tried to apply information theory to such
experiments, there had been no reason to think of the viewer as a channel or
of stimuli as inputs and responses as outputs. If you wanted to estimate the
amount of information transmitted by such a channel, however, it im¬
mediately became apparent that you had to form a matrix with the stimuli as
rows, the responses as columns, and the cells as the frequencies of occur¬
rence of particular stimulus/response combinations. How many rows? How
many columns? In order to estimate the probabilities that were required in
order to compute the transmitted information, answers to such questions
had to be fixed and known. One consequence of trying to apply information
theory to psychology, therefore, was a principled increase in the rigor of
psychological experimentation.
493
494 GEORGE A. MILLER
proximation of English were known, and further research into that question
was unnecessary. By 1960, therefore, psycholinguists had moved on to new
research on the psychological reality of various types of generative gram¬
mars.
* * *
statement can have more than one interpretation and if one meaning is
understood by the hearer and another is intended by the speaker, then there
is a kind of semantic noise in the communication even though the physical
signals might have been transmitted perfectly.
Semantic information never enjoyed the vogue among psychologists that
selective information once did. It is worth reviving here only for the perspec¬
tive it offers on Shannon’s theory.
ENTROPY AND THE
MEASURE OF INFORMATION
Peter Elias
tween the transmitted and received symbols. Its value depends not only on
the channel but on the probabilities with which different input symbols to the
channel are used. If the transmitter chooses to send only one of the possible
input symbols all the time, then the mutual information will, in fact, be zero,
since the a priori and a posteriori distributions and entropies will be the
same. If, instead, the transmitter chooses input symbols from the probability
distribution that maximizes the mutual information, then that maximum
value is called the capacity of the channel and is usually denoted by C and
measured in bits per channel use.
Shannon’s second result can now be stated. It says, roughly, that if a
communications channel has capacity C bits per channel use (in the sense
just defined), then for any positive e,
The first of these results was revolutionary when it was discovered. Its
import is that lack of reliability in a channel does not limit the reliability of
the information that can be received over the channel: It limits only the rate
at which information of arbitrarily great reliability can be received. And
unlike the result of repeating each message an arbitrary number of times to
obtain arbitrarily reliable reception, which gives arbitrary reliability only at
zero rate, Shannon’s second result says that the rate at which information of
arbitrarily great reliability can be received is, in fact, as large as the max¬
imum rate at which unreliable information can be received.
Together, Shannon’s two results allow the solution, in principle, of any
problem of matching a message source to a communications channel, noisy
or noiseless, in two stages. First, take the message source with entropy H
and encode sequences of its messages into sequences of binary digits. Then,
encode each sequence of binary digits into a sequence of channel-input
symbols, and send it over the channel at the allowed rate, decoding the
received noisy sequence into the input binary-digit sequence with probabil¬
ity near 1 and then decoding the input binary-digit sequence correctly into
the original message sequence.
The net effect of these two results is to justify the use of a probabilistic
model of communications for systems in which coding and decoding of
messages is allowed and to demonstrate that in such a model, the entropy H
500 PETER ELIAS
Elliott W. Montroli
Myron Tribus has clearly surveyed the evolution and numerous applications
of information theory and maximum-entropy formalism. I elaborate on the
theme by first making a few historical remarks on the early appearance of the
expression
(1)
j
HISTORICAL REMARKS
I have also found that the logarithm of the ratio, which the middle term of a
high power (n) has to any distance from it by an interval /, would be denoted by
a very near approximation, supposing m = \n, by the quantities (de Moivre,
1756, p. 244)
(m + / — z) log (m + / — 1) +
(m — / + |) log (m - / + 1) - 2m log m + log (m -f 1 )/m. (2a)
503
504 ELLIOTT W. MONTROLL
fm + l , (m + l\ m — l
- log - + log (2b)
m \ m j m
p
fW
o
M -1 Wj log Wj — n. (4)
Today, he would maximize — M, which is in the spirit of the entropy func¬
tion [Equation (1)]. The combination wj log Wj represents the analog of the
terms in de Moivre’s expression [Equation (2)] except that Boltzmann’s
terms followed from the multinomial coefficients in Equation (3), while de
Moivre’s came from binomial coefficients.
Boltzmann then let vv, = /(ye), chose e to be small, proceeded to the
continuum (e = 0) limit, let p —» °°, and minimized
S = -k'Yjfj log ft
ing ingenuity. E. T. Jaynes was one of the first to attempt to unify informa¬
tion theory and statistical mechanics. [Jaynes, 1957a, b.]
with
being the space available per car at location of «th car at time t. The traffic
density at the location of car n, pn(t), is the reciprocal of the space per car: p„
= 1 ld„ = number of cars per unit-length. In a freely moving stable stream of
traffic vn{t + A) with A — 1.5 sec is approximately v„(t), and if pc. is bumper-
to-bumper close-packing density at which vn = 0, Equation (10) becomes
Notice that 0 < p„/pc s 1 and that the dimensions of our variables might be
cars per hour for q, cars per mile for p, and miles per hour for v. By
averaging over N cars in a line of traffic, the mean flow rate is proportional to
an entropy function in the variables (p„/pc). [Montroll, 1981.] Thus,
N
-A
Pn
Q — X0pc (14)
=1 o0 ' '°8 o )/
Pc
N
1
= \0pc. log —-1— log X P» '°g P")' (15)
P N ~ n= 1 >
where p is the average traffic density and pn — p„/pN. While the variables p J
pc are not normalized, the set pn is, so that Xpn = 1. The flow rate is a
maximum at a given traffic density if all p„ are identical, so that p„ = p and pn
= VN. Then the term in the bracket in Equation (15) vanishes. If drivers
behave differently from each other (as of course they do), so that some p„
deviate from p, the entropy term in Equation (15) does not achieve its max¬
imum value and the term in the parenthesis is negative, yielding a reduction
in flow rate q.
fortune, the firm of Sears, Roebuck, and Company has provided a rich
legacy of information on this subject in its annual catalogs, which form a
magnificent database of Americana of the past 85 years.
The preparation of the catalogs was a major concern of Sears Roebuck.
Basically, each page was audited to produce its share of the profit. [Emmet
and Jeuck, 1950.] For example, in 1930 the goals set ranged from $5,000 to
$20,000 per page, depending on the responsible merchandising department.
Since the profit that year was $14,300,000 and the catalog ran 1000-1500
pages, the profit per page averaged about $10,000. Many pages reserved a
small space for the tentative introduction of new products. Favorable public
response motivated increased allocation the next year. With declining sales
of an item, its space allocation decreased; sometimes it even disappeared
from the catalog. Various department heads, anxious for raises and promo¬
tions, were very competitive in preparing pages that listed items that they
hoped would outsell those of their colleagues. Robert Herman and I found
the distribution function of prices by year listed in many of the catalogs.
[Herman and Montroll, 1977.] Since prices ranged from a few cents to hun¬
dreds of dollars, we “expanded” the scale of low-cost items and “con¬
tracted” that of higher priced ones by recording the data as the logarithm of
the price (to the base 2) log2f\ Of course, we were aware (as many before us
dating back to Daniel Bernoulli) that log P is psychologically a more impor¬
tant variable than the price itself because we are especially sensitive to
relative price changes (AP)/P — A log P.
Examining the price distribution from many catalogs indicates that, in a
given catalog, the distribution of log2E, (P, being the price of the /th item) is
very close to the normal distribution [Equation (7)]. Important distribution
moments are
N
and
N
The findings for (log2P) and cr for eighteen years appear in Table 1. The
quantity a is the dispersion in the logarithm of prices relative to their mean
value. The larger the a, the broader the range of prices in the catalog.
The variation in (log2P) over the years reflects changes in cost of living
through the twentieth century. Catalog prices changed in two manners: (1)
by the change in price of an invariant item such as a clothespin, a 1910
specimen being indistinguishable from one in 1940; and (2) by the change in
the nature or quality of the item listed to reflect evolving technology and
varying public taste. Many interesting deductions follow from changes in
(log2E), but it is to the third column a of Table 1 that I wish to direct
THE ENTROPY FUNCTION IN COMPLEX SYSTEMS 509
Hence, to second order, from Equation (17), cr2 of one year transforms to
N
The simplest assumption that might be made is that all items inflate in the
same manner, so that Act, = 0 for all i. Then, cr2 would remain completely
constant over the years. Table 1 indicates that this is almost but not pre¬
cisely the case. If we were to assume that A a,- were independent of the price
P,-, then the middle term in Equation (19) would vanish and a2 would increase
every year, which was not observed. If, however, we assume that low-
priced items were inflated more rapidly than high-priced ones, Aa, would
be positive for small P, and negative for large P,, so that the middle term
510 ELLIOTT W. MONTROLL
CONCLUDING REMARKS
retailing operations in the United States in the first half of this century seems
to have evolved in a manner (unknown to the planners) that maximized an
entropy function in distributing the logarithm of the item prices in its catalog,
subject to dispersion in the square of the logarithm of the prices being
specified and kept constant.
Notice that the last two examples have nothing to do with information
theory or thermodynamic entropy. They seem to reflect the manner in which
a group of people, trying to optimize their productivity, have unwittingly
fallen into a behavior pattern that may be described by an appropriate en¬
tropy function. The diversity of examples of sociotechnical systems previ¬
ously listed suggests that an astute observer should be able to find further
examples of the entropy function in the complex systems generated by peo¬
ple who wish to optimize their productivity.
■
ENTROPY, PROBABILITY,
AND COMMUNICATION
Myron Tribus
The main idea of my paper was that the entropy concept had been used not
only as Shannon originally intended but in novel ways unforseen by him.
Four distinct uses were detected in the literature:
The reviewers have added specific examples of some of these ideas and
because of their specialized expertise and experience, have given us special
insights. In each case, what makes Shannon’s contributions so significant is
the casting of old questions in a new and more interesting light.
I especially appreciate Donald MacKay’s remark, “every unit of missing
information . . . means that a corresponding amount of energy cannot be
pinned down to do work.” The essergy concept, derived in the paper, dem¬
onstrates just that. MacKay also brings out the importance of the constant
K, which in the case of essergy is set equal to A.T0, just as he seems to
suggest.
Although the editors have given me more space for a rejoinder, I do
believe that in view of the reviewers’ remarks, it is better to stop here and
not increase the ratio of words to ideas.
513
SECTION 8
ON SYSTEM THEORY AND
ITS RELEVANCE TO PROBLEMS
IN INFORMATION SCIENCE
Hassan Mortazavian
Research for this paper was carried out while the author was on the staff of the Institut National
de Recherche en Informatique et en Automatique, Paris, France, and he wishes to acknowledge
the many useful comments received from his colleagues at INRIA.
517
518 HASSAN MORTAZAVIAN
member of more than one pair simply because he may have more than one
brother. Imagine now that we relate men in terms of brotherhood, but we
also want to distinguish between older brothers and younger brothers, ex¬
cluding those who are twins. This amounts to first partitioning the elements
of the set into two subsets, one consisting of all those men who are another
man’s older brother and the other consisting of all those who are another
man’s younger brother. Note that the same man then can be a member of
both subsets, in that he may have both an older brother and a younger
brother, hence, being also both an older and a younger brother himself.
Therefore, membership in the two subsets need not be exclusive. After this
partitioning is done, we relate every element of one subset to every element
of the other subset with respect to the relation of brotherhood if such a
relation exists between the two elements of concern. If the first subset
consists of older brothers and the second subset of younger brothers and we
choose the first element of every pair from the first subset and the second
element from the second subset, what we get is a larger set consisting of
what is referred to as a set of ordered pairs.
We need to define one more operation called the Cartesian product of two
sets and denoted by x —the customary sign for multiplication. By the Carte¬
sian product of two sets we mean a set consisting of all ordered pairs whose
elements are chosen from the first and the second set, respectively. To be
more precise, consider the two sets U = {father, mother} and Y = {son,
daughter}. Then the Cartesian product of U and Y is constructed as follows:
Note that the first element of every ordered pair of the product set belongs to
the first set and the second element of every ordered pair of the product set
belongs to the second; thus an ordering between elements is established.
A set S is said to be contained in a set T if every element belonging to S
also belongs to T, and then one writes S C T. Conversely, we may say that T
contains S.
These primitive concepts are sufficient for an elementary definition of the
concept of system, meaning a relation defined on a family of sets.
Consider a simple example adopted from Casti with some modifications.
[Casti, 1979.] Let U be a set of consumer items and Y a set of service
facilities. For example,
Thus,
R y1 Y2 Y3 Y4
X\ 1 0 0 0
*2 1 0 0 0
*3 0 0 0 1
x4 0 1 0 0
branch of mathematics.) The choice between the two directions is not dic¬
tated by subject matter, but by the inclination of the investigator, the pur¬
poses of the research, and the state of development of a particular type of
inquiry.
A definition of system is, as a rule, needed only when systems are pursued
in the analytical direction. For merely constructive characterization, specific
definitions of particular types of systems can be provided that suffice for all
practical purposes. We should distinguish among:
“Does system theory deal with the real world?’’ To this obvious question the
answer is no—that’s quite simple. System theory studies models and does not
accept responsibility for the accuracy or relevance of these models. That head¬
ache is reserved for the practitioners, and it should not be exaggerated. (Kal¬
man, 1980, p. 7.)
This statement means that system theory deals with certain universal
mathematical questions about (abstract or mathematical) models and as such
is independent of particular fields of application. It contributes to other
sciences by providing techniques for generating a class of models for any
well-defined system at hand, based on data resulting from experiment, ob¬
servation, and so forth. But it does not specify which model out of a class of
models is the model for a particular concrete system.
In this sense, system theory is a theory that develops mathematical mod¬
els for various abstract (mathematical) systems. Examples of such systems
or system-types are: linear versus nonlinear systems, continuous versus
discrete systems, deterministic versus stochastic systems, and so torth.
These abstract systems can properly be defined only by appropriate mathe¬
matical tools. Nevertheless, I shall try to offer nonmathematical illustra¬
tions—at the risk of some inaccuracy.
To illustrate what a linear system is, consider the trajectory of the flight of
522 HASSAN MORTAZAVIAN
an airplane. The plane is our system, and we assume that at every point of
time there are only two controls or influences exerted on it: the wind and the
control exerted by the pilot. Each one of these two inputs causes a particular
change in the trajectory of the flight—a particular output. Now, this system
is said to be linear if and only if the output corresponding to any linear
combination of these two inputs is a linear combination of the outputs corre¬
sponding to those inputs. Otherwise, the system is said to be nonlinear.
Roughly speaking, by a linear combination of two variables, we mean a
combination performed only by addition (subtraction) and multiplication by
a Scalar. Note that according to this definition, it is not true that a system is
nonlinear if it contains a nonlinear element. A linear system may indeed
contain nonlinear elements.
A system is said to be continuous if its variables change continuously over
time. A system is discrete if its variables change only at certain specific
points in time. A car traveling on a road is an example of a continuous
system; the number of people entering a shopping center is an example of a
discrete system.
In computing control schemes for real-world continuous-time systems, in
practice, we are forced to treat them, in some sense, as discrete-time sys¬
tems. There are essentially two reasons for this: (1) We can make only a
finite number of measurements of the output values in any time interval. (2)
In computing an appropriate sequence of inputs (controls or decisions) to be
applied to a system to control its behavior in some desired way, we generally
use a digital computer, which is a discrete-time system, and can generate
only distinct control signals at discrete times. The practical necessity of
dealing with continuous-time systems as if they were discrete-time systems
raises important questions about the treatment of sampled-data systems,
that is, systems that should be considered as continuous-time but with which
we can interact only at discrete times when samples of outputs are gathered
and a sequence of discrete controls based on the sampled outputs is applied.
Before providing definitions for deterministic and stochastic systems, I
ought to say a few words about determinism and causality. The word deter¬
mined has often been used in an ambiguous way. In one sense, a quantity—
say the output of a system—is determined when it is measured. In another
sense, an event—say the occurrence of a specific sequence of outputs—is
determined when it is caused. Although there are many reasons to believe in
some principle of causality, it is not altogether easy to define the notions of
cause and causal laws. It must be made clear, however, that what distin¬
guishes deterministic systems from stochastic systems is not the assumption
of causal relation between inputs and outputs or past and future behavior in
one system to the exclusion of the other. Both deterministic and stochastic
systems may or may not be considered as causal (although disbelief in cau¬
sality causes many conceptual difficulties). Deterministic systems are deter¬
mined, and stochastic systems are not determined, in the first sense, which
is independent of the notion of causality.
SYSTEM THEORY AND ITS RELEVANCE TO INFORMATION SCIENCE PROBLEMS 523
of some authors, tends to deal with more specific questions about more
specific types of systems (e.g., dynamical systems, automata, control sys¬
tems, game-theoretic systems, etc.). The main reason for adopting the name
general system theory has been to emphasize the point that system theory
should be considered as a general theory, not as a theory of specific types of
systems, for instance, control systems. In my view, the term system theory
is comprehensive enough to capture all aspects of the field.
David Berlinski has criticized Mesarovic’s definition of a system as a
relation defined on a set of objects on the grounds that this definition col¬
lapses the notion of a system into the notions of a sequence and a set.
[Berlinski, 1976, pp. 28-29.] My response to Berlinski’s criticism is that
almost every abstract mathematical concept collapses into a few elementary
notions such as sets, relations, points, and so on. The notion of a graph, for
example, simply collapses to the notion of points and edges (or elements and
relations between them), and yet graph theory is a powerful mathematical
theory with a wide range of applications. An important advantage of this
definition of system is that it stresses the structural characteristics of phe¬
nomena.
In the second part of this paper, I discussed the concept of system; now, I
shall examine the essential questions of system theory, in particular: (1)
What is system theory? (2) How and why did it emerge as a new discipline?
(3) Why do we need to learn about its findings?
The terms system, systems approach, system analysis, system science,
and so on, are so broadly used, and occasionally misused, that outsiders
sometimes wonder what we really mean by system theory. Moreover, we
often see systems engineering, general system theory, mathematical system
theory, and so on. These terms all refer to various aspects of the same
discipline, to wit, system theory.
System theory does not aim at unifying all or special types of scientific
disciplines. It does, however, provide a methodological base for solving a
number of problems common to various disciplines. Roughly speaking, sys¬
tem theory is the theory of phenomena that are “system-bound,” “system-
dependent,” or “system-related.” By this, I mean those phenomena that
depend on the conditions of the environment in which they exist, interact
with this environment, and thus cannot properly be studied in isolation.
Let us consider some examples of such system-dependent phenomena.
The rate of growth of population in a country, for example, generally de¬
pends on such factors as average income of individuals, income-distribution
patterns, age distribution of population, distribution of male and female
populations, level of education, and so on. Therefore, a model of population
SYSTEM THEORY AND ITS RELEVANCE TO INFORMATION SCIENCE PROBLEMS 525
growth for a country cannot be developed unless we also consider the rele¬
vant variables of socioeconomic environment in that country. Or, consider
the problem of choice of an optimal energy policy for a nation. Such an
optimal policy, if it exists, obviously depends on a number of national as
well as global economic, political, environmental, natural, and other factors
that should be specified for every particular nation.
All such phenomena are system-dependent, and the aim of system theory
is to develop methods that can be used in analyzing, understanding, ex¬
plaining, modeling, controlling, and predicting such system-dependent phe¬
nomena. But techniques of system theory are general in nature, not related
by any means to the subject matter of specific disciplines. System theory
aims at providing a common complement of methods that deal with such
concepts as dynamics, stability, complexity, catastrophe, hierarchy, struc¬
ture, and so on.
Dynamical systems are those that change their state over time. If the state
of a dynamical system changes in the same way over time, it is called time-
invariant, and if the way in which the state changes varies over time, it is
called time-varying. Consider some specific examples: a traveling space
vehicle, a home-heating system, an operating computer, a growing popula¬
tion, the flow of traffic on a road, genetic transference, ecological decay, and
so forth. While each one of these examples refers to a specific dynamic
behavior, the general concept of dynamics transcends all such particular
situations. Countless examples of such dynamic situations in various areas
can be supplied, but they can all be represented by a small number of
mathematical models. System theory is concerned (among other things) with
specifying all such general mathematical models that can represent dynam¬
ics in various situations. This is done by exploring that part of mathematics
that is suitable for modeling dynamic behavior. It turns out that at least the
elementary aspects of smooth dynamical systems can be almost completely
studied by differential and difference equations—the first used when time is
considered to be continuous, the second when time is considered to be
discrete. The velocity of a car, or the flow of water into a tank, changes
continuously; therefore, velocity and flow are modeled in continuous time,
hence, by differential equations that relate the rate of change of a variable to
its level at every point in time. The number of people entering a shopping
center, on the other hand, changes discretely; therefore, it is modeled in
discrete time, hence, by difference equations that relate the value of a vari¬
able at one time to values at adjacent times. That part of system theory that
provides models for such dynamical phenomena, using mostly differential
and difference equations, is called the theory of dynamical systems.
[Kalman, Falb, and Arbib, 1969; Brockett, 1971; Mesarovic and Takahara,
1975; and Luenberger, 1979.]
‘Of course, at a higher level of sophistication, the theory of dynamical systems requires a
number of algebraic, geometrical, topological, and catastrophe-theoretic concepts as well.
526 HASSAN MORTAZAVIAN
Although I cannot in the allotted space give more than only the sketchiest
outline, I wish to emphasize that whenever we deal with complex interactive
phenomena, we cannot avoid using the results of system theory, in the same
way that we cannot avoid using the results of physics or chemistry when we
want to study the nature of physical objects.
2 Among the most important of the contributions are: Kalman, Falb, and Arbib [1969], Bensous-
san [1971]; Bensoussan and Lions [1978; 1981]; Bensoussan, Delfour, and Mitter [1982]; Bell¬
man [1971]; Brockett [1971; 1978]; Brunovsky [1970]; Byrnes and Falb [1979]; Fuhrman [1979];
Hazewinkel and Kalman [1976]; Hermann and Martin [1977]; Martin and Hermann [1978];
Kalman [1960a; 19606; 1962; 1965; 1971; 1972; 1976; 1979; 1980; 1981; 1982]; Klir [1969];
Luenberger [1964; 1966; 1971; 1979]; Mesarovic [1964]; Mesarovic, Macko, and Takahara
[1970]; Mesarovic and Takahara [1975]; Pontryagin et al. [1962]; Popov [1972; 1973]; Rosen-
brock [1970]; Rouchaleau and Sontag [1979]; Tannenbaum [1981]; Thom [1972; 1974; 1977;
1980]; Willems [1979]; Wonham [1967; 1974]; Zadeh and Desoer [1963]; Zadeh and Polak [1969].
528 HASSAN MORTAZAVIAN
social sciences; I am simply emphasizing the need for, but want of, appropri¬
ate applications.)
Before proceeding further, I would like to suggest the following definition
for the discipline of system theory:
The true meaning of this definition, and the justification for its adoption, will
become clear in the following pages.
In order to present some of the specific results of system theory, I need to
return occasionally to the discussion of the specification of systems though
this will involve some repetition. I will first take up a particular type of
specification, called input/output description because the system is de¬
scribed in terms of inputs and outputs. If inputs and outputs of a concrete
system are distinguishable, then it can be modeled or represented by an
abstract (mathematical) input/output system.
Following the definition of a system as a relation (or relations) defined on
a set, we may now partition the elements of the set into two object sets, one
consisting of input elements, the other of output elements. In this way, a
constructive definition of a system can be provided. A concrete input/output
physical system is a part of the world isolated naturally or artificially from
the rest of the world and considered as a box in which something (matter,
energy, or information) enters at certain times and that itself puts out some¬
thing at certain times. At each moment of time, the system, be it continuous
or discrete, receives some input and produces some output. When the sys¬
tem produces output discretely, we assume that the output between two
points of time is simply zero. We assume that values of the input are taken
from a specific set, which for every system will be defined depending on the
circumstances. For example, a country’s imports are often restricted by
both governmental regulatory policies and the conditions of the interna¬
tional-trade system. Or, another example, the amount of information a par¬
ticular computer can process in every second is limited to the capacity of its
central-processing unit (CPU). Such restrictions on inputs are quite natural
and related either to bounds on inputs or the nature of the inputs. Thus, we
define a set of permissible or allowable inputs to the system. In the same
way, we define a set of permissible outputs from a system, though with much
milder restrictions.
A philosophical justification for such an input/output description of a
system helps make it clear that the notion of an input/output system is not
simply a practical device. If we consider the system S contained in a specific
domain D of the world with walls around it, then we need somehow to get
inside this domain in order to be able either to control it or observe it. This
remark paraphrases one made by Rene Thom:
SYSTEM THEORY AND ITS RELEVANCE TO INFORMATION SCIENCE PROBLEMS 529
... a system is never completely independent of us, human observers (if it was,
we had better forget about it . . . ); hence the walls of D have to be provided
with windows, through which we may act on the system or, conversely, re¬
ceive information about its inner state. This justifies introducing the notion of
Input and Output of the system, or—in slightly different terms—of control
parameters and of observables. (Thom, 1980, p. 3.)
’For a more detailed discussion, see Luenberger’s very instructive elementary textbook.
[Luenberger, 1979.]
SYSTEM THEORY AND ITS RELEVANCE TO INFORMATION SCIENCE PROBLEMS 533
2. When sufficient behavioral data are not available (the word sufficient
in this context is a well-defined term in realization theory), the class of
constructed realizations among which a model is to be chosen becomes
unreasonably large. Research in this area is in progress.
3. Data-dominated modeling is mathematically precise and safe, but to
some extent, it lacks the transparency of theory-dominated modeling in the
sense that it may not be easy to give concrete interpretation to the relations
contained in the model. Whether or not this is to be considered a serious
difficulty depends partly on the type of problem at hand and partly on the
purpose of the modeling.
(i) There is one and only one parameter set corresponding to any system in the
class.
(ii) The constraints on the parameters are explicit.
5Stochastic realization theory has a rich literature: Kalman [1965]; Rissanen and Kailath [1972];
Faurre [1973]; Akaike [1974]; Picci [1976; 1977]; Faurre et al. [1979]; Van Putten and Van
Schuppen [1979].
SYSTEM THEORY AND ITS RELEVANCE TO INFORMATION SCIENCE PROBLEMS 539
hBy “empirically definable relations,” these authors mean those relations that can be estab¬
lished among objects as a result of measurements performed on them.
SYSTEM THEORY AND ITS RELEVANCE TO INFORMATION SCIENCE PROBLEMS 541
It is high time that we ask ourselves what we, and others, too, mean by
the term information science. By information science, we mean the assem¬
blage of systematic studies aimed at understanding, interpreting, analyzing,
and measuring information; and modeling, organizing, and utilizing the pro¬
cess of transferring information, or more generally, knowledge—be it among
humans, humans and machines, or only among machines. From this per¬
spective, communication theory and the theory of semantic information are
considered only as parts, though significant parts, of the broader discipline
of information science.
Throughout this paper, I have tried to point out the relations between system
theory and information science. It may be helpful if I attempt a synthesis of
my scattered remarks on this question.
An information system in the nonsemantic sense of the term is a special
type of relation or transformation defined on two reference sets: inputs and
outputs. In this case, information generated by the information source may
be considered the input to the system and information derived from the
system the output of the system. The information system can be defined in
this manner if it is both continuous (e.g., nondigital measuring-devices such
as a voltmeter; or devices for measuring pitch and amplitude of a sound
wave, color and brightness of a graphic pattern, field strength of a radio
emission, etc.) or discrete (e.g., a written language, a digital computer, etc.).
Also, as Thom remarked, there is a conceptual relation between an input/
output representation of a system and the requirement to interact with the
outside world (observe the outside world) in an information-theoretic sense.
Indeed, Thom goes as far as relating the apparent time reversibility of physi¬
cal laws with the concept of semantic information. Thom states:
Putting things more abruptly, I would dare to say that the time reversibility of
physical laws is probably no more than the expression of a sociological con¬
straint, namely communication between several observers. For this constraint
is nothing more than the linguistic constraint between members of the same
linguistic community: when people speak the same language, they share the
same semantic universe: because, to the same sentence, they have to put the
same meaning (or at least, approximately the same). In fact, any observer has
to communicate with himself—with his own past. Flence he needs to have the
possibility of comparing his way of looking at the universe at time t\ with the
look he had at time t0< tp, this requires a common standard of description, a
permanent way of parameterizing the states of the world. Hence reversibility
of the dynamics. (Thom, 1977, pp. 193-194.)
from other circles (i.e., its radius) are descriptive and, hence, can provide us
with only descriptive information about the circle. Those properties of a
circle, on the other hand, that are common to all circles and make it distinct
from anything that is not a circle are intrinsic and, hence, can provide us
with structural information about the circle. The set of properties that distin¬
guish circles from noncircles is invariant under changes in descriptive
parameters. A similar argument applies to various classes of systems.
CONCLUSIONS
System theory applies to any branch of science in which we deal with com¬
plex interactive phenomena irrespective of the context in which the problem
may arise. System theory is a mathematical discipline though not entirely a
branch of mathematics. It is the theory of mathematical models and applies
to various concrete situations by providing practitioners with tools and tech¬
niques to develop models for concrete systems. More specifically, if a con¬
crete system is described in terms of a sequence of input/output data, the
task of the system theorist is to provide answers to the following questions:
Answers to these questions are known for certain classes of systems. How¬
ever, we are still far from being able to provide complete answers to all these
and several other questions for all classes of systems.
Let us emphasize, however, that whenever we have to deal with complex
interactive phenomena we cannot avoid using the results of system theory, in
the same way that we cannot avoid using the results of physics or chemistry
when we want to study the nature of physical objects.
SYSTEM THEORY, MATHEMATICS,
AND QUANTIFICATION
Kenneth E. Boulding
ters, and innumerable relations for which there are no mathematical equiva¬
lents. The idea, therefore, that mathematization and quantification are the
only hallmarks of science seems to be quite fallacious. It has done substan¬
tial damage even in the biological sciences, but especially in the social sci¬
ences. I was glad to have my attention called to a remarkable article by B. C.
Brookes entitled “A New Paradigm for Information Science?” that dis¬
cusses these considerations in more detail. [Brookes, 1976.]
In the early days of what is now called the Society for General Systems
Research (of which I was a founding father and first president), I recall
defining a system as anything that was not chaos, and I could hardly blame
Mortazavian if he felt that such a definition was too broad and philosophical
to be very useful. We did, I also recall, define a general system as any
theoretical structure that was of interest to more than one academic disci¬
pline. This, perhaps, is a concept too sociological, especially in light of the
rather accidental nature of the academic disciplines themselves, to have
much prestige. Nevertheless, we cannot avoid the epistemological over¬
tones of the systems concept; if system is equivalent to order (that is, non¬
chaos) or even some useful subset of things, structures, or relations that
exhibit order, the critical question is how we come to perceive, recognize, or
know about the order that presumably exists in the real world. This is nearly
the same question as how do we get order—patterns and structures—inside
our skins, and especially inside our skulls, that represent or map onto corre¬
sponding structures in the real world. The system of human learning is
immensely complex and still very imperfectly understood, but we must
know something about it or we would not be able to transmit the knowledge
structures from one generation to the next, even though we do not know
much about how these knowledge structures are coded inside the body.
In the formation of our knowledge structures, perhaps three processes
can be identified. One is the building of knowledge structure in the nervous
system by the genes, which used to be called instinct. The second is thought,
that is, the internal building up of new knowledge structures by processes
that involve consciousness. A very important element in this is the percep¬
tion of identities—that is, relations that cannot be other than what they are.
Mathematics has a peculiarly important role to play in perceiving identities;
it might almost be defined, indeed, as the sophisticated pursuit of the obvi¬
ous. Mathematical proof is simply the process of finding out what is indeed
obvious, which is not always obvious at first. There is, indeed, a slight rattle
from a skeleton in the closet here, partly from Russell and partly from
Godel, even perhaps partly from some Godot for whom we are still waiting
and who might point out that what is obvious is not always true. It seems
obvious to me that through a point we can draw only one line parallel to
another line not going through the point, but it was not obvious to Riemann
or Einstein. Similarly, it seems obvious that minus minus is plus, but I worry
about the fact that not doing harm is a very different affair from doing good.
It seems obvious, too, that infinity times zero is any finite number we like to
SYSTEM THEORY, MATHEMATICS, AND QUANTIFICATION 549
name though this may involve definitions of both infinity and zero as things
toward which we might travel hopefully but never arrive at, especially as
infinity, and I presume also zero, has more addresses than we used to think.
1 worry though about calculating price indices when the price of a color
television set in 1910 was infinite and the quantity was zero!
The third process in the formation of knowledge structures is experience,
which represents a very complex structure of inputs into and outputs from
the body, certainly involving B.T. information, but also involving much
more than that in terms of structure, for instance, in language or even sense
perception. We suspect that if information input is to change knowledge
structure within the body, it must itself be structured in such a way that it
“fits” into some internal pattern. (I believe strongly in the survival of the
“fitting” rather than of the fittest!) If someone addresses a remark to me in
Chinese, a language with which, I regret, I am unacquainted, it will have a
very different impact from the same remark translated into English, even
though both remarks may contain the same number of bits of B.T. informa¬
tion. Even the expression B.T. information will have a different impact on
my reader in the last sentence than it would have had before reading the
paper.
In the learning process, the impact of current inputs, whatever they are,
depends on the existing structure of the mind, and this depends in no small
measure on the existence of records of the past. We go to a familiar place
and are able to find our way around it, because in our minds there is a
structure (memory) that has recorded not only past inputs but past struc¬
tures of knowledge. The success of science as a knowledge-expander de¬
pends in no small measure on that aspect of the scientific subculture that
consists of keeping careful records of the past, whether this results in maps,
the paths of the planets, national income statistics, or experimental results.
Both observation and experience, the two great sources of the inputs of
science, involve records of the past. Our perception of patterns in the
records of the past is the foundation of both prediction and experiment,
which is usually prediction of some future pattern of small contrived sys¬
tems. Our ability to interpret such records and patterns, however, depends
largely on whether they fit into previously discovered obvious relations,
which are usually, though not always, mathematical in type. I have argued
that the great laws of science are either truisms (mathematical identities) or
near truisms, relations that we are almost sure have to be this way, which
might rank as axioms. Much empirical work is devoted not to finding empir¬
ical laws, which are always unreliable, but in spotting where, in the empir¬
ical fields, the basic truisms apply. A good example is the principle of con¬
servation, which simply says that if there is a fixed quantity of anything, any
increase in one place or form must be offset by decreases in other places or
forms. The empirical question is “What is there a fixed quantity of? En¬
ergy? Matter? Money? And so on. A related identity is what I have called the
“bathtub theorem”—that the increase in anything in a given period is equal
550 KENNETH E. BOULDINC
Richard N. Langlois
1 “I hasten to add that system theory and mathematical system theory refer to one and the same
thing. There is no such thing as a nonmathematical system theory.’- (Mortazavian in his paper
in this volume.)
2I touch briefly on these issues in my own paper in this volume.
551
552 RICHARD N. LANGLOIS
3The literature here is voluminous and wide-ranging. For an intelligent defense of the systems
view, see C. West Churchman [1968b] and for a hysterical one, see Simon Ramo [1969]. For
representative critiques, see Laurence H. Tribe [ 1972] from a philosophical perspective and Ida
Hoos [1972] from an empirical point of view.
CAN MATHEMATICAL SYSTEMS BE CONCRETE? 553
method is simply to describe the system with which he or she deals. The
scientist is thus a handmaiden to the great generalist, a kind of graduate
research assistant to the system theorist, whose task it is merely to provide
the right description of the concrete system. It is the system theorist who
performs the real scientific function, spinning the mathematical wheels to
identify the correct equivalence-class of models, which is then given back
into the grateful hands of the scientist.
Fortunately or unfortunately, however, we do not live in this simple
Platonic world. Complex phenomena of nature, mind, and society (concrete
systems) do not themselves possess variables; only mathematical models
have variables. The job of the scientist is not to describe such phenomena in
some nonproblematical fashion; rather, the scientist’s task is one of active
interpretation, mental construction. And mathematical models are often—
but by no means always—useful tools in this creative process. Thus, the
mathematical system theorist, who studies the logic of operation of such
models, retains an important role in the scientific endeavor—even if it is the
same role mathematics has always played in science rather than the sup¬
posedly new and distinctly baronial role the mathematical system theorist
would wish to be assigned.
SYSTEM THEORY AND
INFORMATION SCIENCE
Further Considerations
Richard Mattessich
The claim that system theory and mathematical system theory refer to one
and the same thing, and the attempt to identify the broad discipline of system
theory with one and only one of its many subareas, is bound to lead to
confusion. Thus, it is important to realize that whenever Mortazavian, in his
paper in this volume, speaks of system theory, he limits himself and his
audience to the relatively narrow area of mathematical system theory. Al¬
though authors should be given freedom to choose their own terminology (or
that of their circle), that freedom must be curtailed whenever convention is
deeply enough rooted in a different meaning—otherwise misunderstanding
is inevitable. But there might be more behind this narrow definition. Is it,
perhaps, an attempt to declare everything beyond the mathematical part in
system theory as trumpery or unscholarly endeavor?
Additional support from the Social Sciences and Humanities Research Council of Canada is
gratefully acknowledged.
555
556 RICHARD MATTESSICH
.
4 A thorough study of the emergence and evolution of system theory
offers clear evidence that the original impetus for this novel viewpoint arose
from the inadequacy of the traditional atomistic and reductionistic approach
when dealing with such factual phenomena as emergent (holistic) properties,
environmental interactions, feedback information, and so forth. [Bogdanov,
1913 and 1922; Bertalanffy, 1928; 1950; 1968.] This inadequacy was hardly
bothersome in physics but was strongly felt in the life sciences; and it is no
coincidence that of the two pioneers of system theory, Bogdanov and Ber¬
talanffy, one was a physician-philosopher, the other a biologist. The fact that
both of them envisaged this new area as a superscience instead of mere
methodology is unfortunate but affords no justification for mathematicians
to usurp system theory and claim it entirely for themselves or deprive it of its
empirical and methodological content.
.
5 The attempt to banish the factual and normative-methodological es¬
sence of system theory stands in crass contradiction to the present state of
the sciences. System methodology has been so widely accepted that system
theories not only evolve in operations research and applied mathematics,
but emerge, in one form or another, in accounting theory, organization
theory, and economics, the management sciences, electronic data process¬
ing, and engineering, and even in epistemology and ontology, of which
Reseller’s and Bunge’s rigorous and integrative work are excellent exam¬
ples. [Rescher, 1979; Bunge, 1979a.]
6. Finally, the system approach, due to its goal orientation, seems to be
particularly suitable as a basis for formulating instrumental hypotheses
(means/end relations), which are of special importance in the applied and
social sciences. This entails the emergence of novel methodological and
other epistemological problems, which I examined in a previously men¬
tioned work. [Mattessich, 1978; 1974a; 1982.]
theory is already well supported, while in the second case, it is either (1) a
provisional hypothesis or (2) a mere conjecture that is not expected ever to
fulfill the function of a scientific law.
Furthermore, the paragraphs devoted by Mortazavian to the genealogy of
system theory leave something to be desired. Although Bertalanffy is briefly
referred to, no mention is made of Bogdanov’s more original work [1913 and
1922] nor of the contributions to this area by Ackoff [1963; 1971], Ackoff and
Emery [1972], Bunge [1979a], Churchman [1964; 1968a; 19686; 1971; 1979],
Mattessich [1974a; 1978], J. G. Miller [1978], Simon [1969 and 1981], and
many others. Indeed, even within the subarea of mathematical system
theory, Mortazavian’s references seem to be somewhat lopsided, since his
bibliography contains only a single article from the important journal Mathe¬
matical Systems Theory.
All this is neither meant to deprecate contributions made by Mortazavian
nor to deny the usefulness of mathematical system models. On the contrary,
Bunge’s ontological investigations, for example, provide a comprehensive
axiomatic system-framework and thus acknowledge the importance of math¬
ematical formulations in system theory. But Bunge, choosing the term sys-
temics—as he holds that there is no single system theory but many of
them—emphasizes that
Systemics has two related motivations, one cognitive and one practical. The
cognitive or theoretical rationale of systemics is, of course, the wish to dis¬
cover similarities among systems of all kinds despite their specific differences.
. . . The practical motivation for systemics is the need to cope with the huge
and many-sided system characteristics of industrial societies. . . . (Bunge,
1979a, p. 1.)
The research that was done to design computer time-sharing systems is a good
example of the study of computer behavior as an empirical phenomenon. . . .
To understand them, the systems had to be constructed, and their behavior
observed. . . . Here again theoretical analysis must be accompanied by large
amounts of experimental work .... (Simon, 1969 and 1981, pp. 24-25.)
two different connected things.” (Bunge, 1979a, p. 6.) Perhaps I should note
that, for Bunge, a thing is always a concrete object and a connection is more
than a mere relation, since “two things are connected (or coupled or linked
or bonded) if at least one of them acts on the other.”
The somewhat narrow treatment by Mortazavian not only misses the empir¬
ical and methodological aspects of system theory, but equally neglects the
important area of information economics. Such an omission appears to be
particularly critical wherever an overall synthesis between system theory
and information science is invoked, as it is by Mortazavian. His broad
definition of information science gives no indication of excluding informa¬
tion economics and thus presumably would encompass it. But as Mortaza¬
vian neither discusses nor even mentions this important area, some pertinent
remarks would be due in this paper. I shall, however, deal with information
economics in connection with Richard Langlois s paper, System Theory,
Knowledge, and the Social Sciences,” elsewhere in this volume.
PROBLEMS OF SYSTEMS THEORY
C. West Churchman
The lead papers in Sections 8 and 9, both of them on the topic of systems or
the systems approach, illustrate a phenomenon common to academics who
write from the point of view of different disciplines, namely, the ignoring of
large hunks of literature on the same topic. I shall add to the confusion and, I
hope, to the enlightenment by discussing still another systems approach,
which both authors have largely ignored. An approach is toward something,
I think (though in the case of the two papers, it seems as though approach is
equivalent to method). The systems approach I want to discuss is an ap¬
proach to planning.
Planning, a very old human activity, means the attempt to understand
enough about a human social system to be able to formulate policies that, if
implemented, will lead to an improvement in the system. This is not a
definition of planning in a reductionist sense, since it uses such words as
understand, policies, and system, which are at least as difficult to define as
the word planning itself. But the description does indicate that we, the
planners, are in the business of trying to help policymakers primarily, and
only secondarily do we strive to excite the intellectual curiosity of people
working in various academic disciplines.
The systems approach to planning is old. I date its first writings back to
the I Ching (say 2000 B.C.),1 which attempted to classify situations in which
humans find themselves and to suggest attitudes about each of them that
could create wisdom in action. Many other ancient and modern books may,
or should, be cited, that I believe say some important things about planning
and a systems approach to it.2 The reader may notice that these books are
written from an extraordinary variety of points of view and mathematics and
economics are no more dominating than psychology, biology, law, psycho¬
analysis, political science, ethics, and epistemology.
‘Compare Richard Langlois’s estimate: “Economists . . . were the first systems theorists.”
2The list should not omit the following: the Bhagavad-Gita, Plato, Aristotle, Lucretius, the New
Testament, St. Augustinus, Hobbes, Spinoza, Kant, Bentham, Edgar Singer [1948], von Ber-
talanffy [1968], Emery [1969], Habermas [1970], Van Gigch and Hill [1971], Beer [1972], Ackoff
and Emery [1972], Hoos [1972], Bateson [1972], Meadows et al. [1972], J. G. Miller [1978],
Jantsch [1980], Checkland [1981],
561
562 C. WEST CHURCHMAN
bernetics, queuing theory, and so forth; so there are as many styles of the
systems approach as there are practitioners. For the reader interested in
applications, there is the journal Interfaces, which publishes applications
together with an independent evaluation establishing that the application has
worked. For a compendium of theoretical styles, see Peter Checkland’s
recent book Systems Thinking, Systems Practice. [Checkland, 1981.]
In the remainder of this comment, I shall consider two quite related topics
that also appear in the papers by Richard Langlois and Flassan Mortazavian,
namely, systems information and the relation between systems thinking and
the logic of classes.
There is a fascinating aspect of the logic of planning for action, namely,
that if someone does x, it follows that he or she does not do non-x, assuming
that the descriptors of x are complete (i.e., that within a time span, all the
relevant things he or she does have been enumerated). This statement is not
really an axiom of systems planning because it is very close to being a
tautology. What makes it fascinating is an axiom of all systems planning
today:
Axiom 1. The class non-x not only is nonempty, but also contains a
nonempty subset of actions that could have been taken by the decision¬
maker in the time span.
We even have a technical phrase for this nonempty set: “lost opportuni¬
ties.” The etymology of to decide is very appropriate, since the verb comes
from the Latin decidere, which means to cut, or to cut off. Hence, if action x
is taken during a time span, then the actor (decisionmaker) cuts off a
nonempty set of lost opportunities—forever. That one or more of this set
could be chosen later on does not change the fatality of the lost opportunity,
especially in that large number of cases where timing is crucial. (In chess, for
example, we cannot say to our opponent, “I failed to take your queen with
my pawn on the last move, so now I would like to do it on this move even
though you have moved your queen away.”)
Why is the class of lost opportunities so fascinating to the systems plan¬
ner? Because the characteristics of this class contain a most vital piece of
information, namely, the cost of doing x rather than doing something in the
class of lost opportunities. Now I need another axiom, perhaps not so self-
evident as the previous one:
Axiom 2. For any action x taken over a specific time span, the class
of lost opportunities can be ordered in terms of their value to a class oj
human beings.
Here, I need to point out that this ordering is no simple matter. In texts on
decision-making, it is often assumed that an alternative action is like pulling
the lever on a slot machine, where the action runs itself and we simply await
564 C. WEST CHURCHMAN
the outcome. But if the action includes such matters as investing money,
firing or hiring personnel, or selling property, we have to know how to do
these things well or even optimally. Hence, in assessing the value of a lost
opportunity, we have to understand how the opportunity can best be
managed.
This is a crucial point in the systems approach to planning because in
trying to determine how best to manage a lost opportunity, we are taken out
of the original problem statement into other systems.
Finally, if I can assume that the value scale that orders the lost opportuni¬
ties is additive, then
This last axiom need not be so strongly stated; instead, we can use the more
qualitative prescription: The planner must take full account of opportunity
costs.
So much for formalism. Now, I would like to ask an epistemological
question: “How do we come to know opportunity costs?” They are vital
pieces of information for all planners who use the systems approach, and,
hence (assuming my axioms hold), such planners should have an appropriate
epistemology for them.
In the first place, it is clear that determining an opportunity cost is not
empirical. The lost opportunities are all “counterfactual” (if y had been
done instead of jc, then . . . ). Besides, it would be a hopelessly impractical
task to test the consequences of each of the members of the correctly man¬
aged lost opportunities even if we knew what correctly managed meant. If
this comment is valid, then it appears that what is called information theory,
which assumes an empirical base for the information, is largely irrelevant as
far as opportunity costs are concerned.
Rationalism would appear to be a more attractive epistemology with ref¬
erence to Axiom 3. Suppose that the planner has a sufficiently strong theory
to predict the cost of every possible action the decisionmaker could take
over the time span. This, indeed, seems to be the case in linear program¬
ming. The geometry of linear programming is designed to map in multi¬
dimensional space all the possible costs for each set of possible actions.
Actually, we do not normally find any discussion of opportunity costs in the
literature of linear programming, perhaps because the authors think that
their measurement is more or less automatically carried out.
In practice, the opportunity costs are all there, hidden by the method of
estimating the parameters; for example, measures of the amount that a unit
of an activity contributes to the measure of total cost. This very often is
PROBLEMS OF SYSTEMS THEORY 565
estimated empirically but without determining whether the unit that contrib¬
utes to total cost was optimally managed or whether increasing or decreasing
any particular output might lead to considerable organizational disruption.
The epistemological problem of estimating opportunity costs is more eas¬
ily recognized if we turn to a very familiar system, inventory. Suppose my
job is that of a purchasing agent for a retail store. How many white shirts,
size 151/2, 33, of a certain make should I order at one time? If I order too few,
I run up ordering costs. If I purchase too many, I shall face obsolescence
costs, and more to the point, I shall reduce liquid capital. The shirts sit on
the shelf for a number of days and do not produce cash. What could best be
done with the cash they cost me is the opportunity cost of capital, and we do
try to use this number in estimating optimal order quantities. In other words,
the correctly managed lost opportunity is a matter of correctly using the
liquid capital.
But how do we obtain the estimate? Many textbooks on inventory control
tell the student to use the interest that the cash could earn. Aside from the
fact that there is no such thing as the interest in today’s money markets, the
suggestion is absurd at the policy level. If the best the retail store can do with
liquid capital is to invest it in interest-earning accounts, then the store should
liquify all its assets and set them to earning interest.
The question “What is the opportunity cost of holding inventory?” de¬
pends on determining the answer to another question, “What is the best use
of liquid capital?” which depends on the answer to still another question,
“What is the best use of any capital?” which asks for a total financial plan of
the company.
I hope the reader can begin to see why systems information does not
correspond to empirical information at all and why its acquisition is epis¬
temologically mysterious. Far from being able to reduce the question of the
appropriate opportunity cost to simpler questions, we find that it expands
into questions about the larger system. The following is not an axiom, but
more like a theorem:
The more we investigate how a system should work, the larger the
system becomes; or
Under investigation, the boundaries of systems keep breaking out¬
wards; or
The essence of the systems approach is the interconnection between
systems.
they influence what happens. Like any other profession, we make the best
guesses we know how to, and we have plenty of hope.
In closing, I would like to make two comments, one more or less technical
and relevant to the two papers, the other more general and germane to world
systems. The technical question deals with the manner in which systems
should be depicted. Here, it seems common sense to say that systems are
made up of parts (components, departments, and the like) and that the
systems-planning and management problem is to determine how the parts
should work together for the good of the whole system (e.g., its goals). All
this common sense suggests that the logic of parts and wholes is the appro¬
priate beginning of systems theory: Belongs to is reflexive, nonsymmetric,
transitive, and so forth. Also, if a belongs to b and b belongs to a, then a and
b are the same.
However, if I want to justify an examination of part b when my interest is
in part a, I would like to define is a part of somewhat as follows: b is a part of
a if the way in which b operates influences strongly the performance of a.
Please understand that this definition is epistemological and might more
reasonably be posed as: Investigation of b is part of the investigation of a if
. . . . But now, the part/whole logic will appear with an overall symmetry.
So my technical point is that the logic of classes is not the basis of systems
theory, as Mortazavian seems to suggest.
My other point is based on the idea that we need to concentrate on
systems interconnections. I think the point was made very well in the first
book on world modeling, The Limits to Growth. [Meadows et al., 1972.] My
chief interest in the past several years is in the social diseases called militar¬
ism and malnutrition. I and my colleagues at Berkeley have been trying to
see in what sense these are related to each other. Seeing the relation does
not necessarily solve every problem, but it may lead to more enlightenment.
Here is the way we have been thinking about the interconnection:
Mihajlo D. Mesarovic
Hassan Mortazavian
that is, modeling based on data not laws, Langlois finds that it “implied a
naive faith in data long ago renounced—with good reason—by the philoso¬
phy of science.” To be sure, observation alone, and thus modeling based on
data alone, can never discover new laws or new theories. New laws can only
be discovered by new theoretical analysis and new experimentation. What I
said about the data-dominated approach to modeling does not amount to a
“naive faith in data.”
Langlois writes that Kalman’s view of the world and to some extent my
own view “serves in the quest to unify all sciences under the banner of
system theory . ...” I have said exactly the opposite in my paper when I
wrote in the beginning of the third section, “system theory does not aim at
unifying all or special types of scientific disciplines.”
Contrary to Langlois’s interpretation of my views, I do not regard the
scientist as a handmaiden or “a kind of graduate research assistant to the
system theorist.” System theory is a science like any other science, and in
science, there is no question of inferiority or superiority.
In a footnote, Langlois tried to clarify the special meaning of the term
observability adopted by mathematical-system theorists. Langlois is right
when he says that the notion of observability has nothing to do “with the
phenomenology of observing a system.” I said this myself in my paper:
“Observability of a system . . . should not be mistaken for the ordinary
process of observation.” However, the rest of Langlois’s footnote is partly
misleading, partly wrong. The point is not that we “know all about a sys¬
tem,” or find out “what the (known) state variables are doing.” The ques¬
tion is whether we can measure the values of the state variables of a given
system at every instant of time within a time interval from the measurement
of its outputs.
I now turn to comments by Richard Mattessich. He warns the reader that
whenever I speak of system theory I limit myself to “the relatively narrow
area of mathematical system theory.” I have argued in my paper that non-
mathematical system theory does not and cannot exist and, therefore, that I
take system theory and mathematical system theory to be identical. The
term mathematical system theory was introduced, I think, for two reasons:
One, to distinguish true system theory, which must be mathematical, from
works that were not mathematical and yet were called system theory by
their producers. Second, to distinguish the part of system-theoretic research
that used more sophisticated mathematical tools from parts that used less
sophisticated and sometimes poor mathematics. I cannot think of any sys¬
tem problem that can be solved without mathematics. If such a problem
exists, it must be a trivial “nonproblem.” I hasten to add two remarks: (1)
System-theoretic ideas did not arise from mathematics, but rather from the
practical necessity of dealing with complex concrete phenomena; (2) it is
possible, and indeed quite probable, in any science to miss the real problem
and fall into the trap of too much mathematical sophistication.
Mattessich criticized me for not having mentioned the works of Ackoff
and Emery, Boulding, Bunge, Churchman, James G. Miller, and Herbert
SYSTEM THEORY VERSUS SYSTEM PHILOSOPHY 577
'The scientific contribution of these authors, however, is by no means limited to this domain.
SECTION 9
SYSTEMS THEORY,
KNOWLEDGE,
AND THE SOCIAL SCIENCES
Richard N. Langlois
This research was supported in part by the Division of Information Science and Technology of
the National Science Foundation under grant IST-8110537. Note: Portions of this paper ap¬
peared under the title “Systems Theory and the Meaning of Information” in “Perspectives on
Systems Methodology and Information Research,” a supplement to the Journal of the American
Society for Information Science, vol. 33 (November 1982), pp. 395-399.
'See Hassan Mortazavian’s contribution to this volume.
581
582 RICHARD N. LANCLOIS
grasping all there is to know about the whole—society. In other words, the
properties of the whole can be deduced from the properties of the parts; or to
put it in more familiar (and more naive) terms, the whole is just the aggregate
or sum of the parts. This view is to be contrasted with methodological holism
(or, sometimes, collectivism), which insists that wholes possess “emergent”
properties that cannot be derived from the properties of constituent parts.
To the holist, the whole is greater than the sum of its parts.
This is a familiar story. But, as is often the case with familiar stories, it is
almost entirely wrong. There may well be some writers who espouse in
principle this sort of naive methodological individualism. But no one can
actually put such a view into practice. Consider the case of a neoclassical
economist analyzing the effect of a tax suddenly placed on a certain com¬
modity. In proper individualist fashion, he or she will instantly begin to
model the choice problem faced by a representative economic agent, discov¬
ering, as always, that the individual will most likely consume less of the
commodity than before. But this finding about individual demand tells the
economist nothing about the whole—total demand—until he or she adds a
global fact: that total demand is the sum of individual demand. In this (triv¬
ial) case, the whole is just the sum of the parts; but even here, the whole
could not be deduced from the parts, since the relation among them—
addition—is not logically contained in the individual-choice model itself.
Moreover, there is nothing sacred about addition, and methodological indi¬
vidualists are quite willing to specify very different sorts of relations among
the parts. In economics, for example, the aggregate result is very often
exactly the opposite of what we would have expected from considering
individual behavior alone.2
Far from denying the importance of emergent phenomena, the economist
and philosopher Friedrich A. Hayek, widely (and correctly) cited as an
archproponent of methodological individualism, reminds us that the entire
objective of the social sciences is to explain how the behavior of individuals
leads to orderly patterns and institutions that none had consciously
planned—to explain, in other words, the emergent results of individual ac¬
tion. [Hayek, 1979, pp. 146-147.]
In sophisticated discourse, the question of emergent properties is in no
way an issue between individualist and holist. Both agree that social phe¬
nomena must often be considered emergent wholes whose behavior cannot
3 “The overall order of actions in a group is in two respects more than the totality of regularities
observable in the actions of the individuals and cannot be wholly reduced to them. It is not so
only in the trivial sense in which a whole is more than the mere sum of its parts but presupposes
also that these elements are related to each other in a particular manner. It is more also because
the existence of those relations which are essential for the existence of the whole cannot be
accounted for wholly by the interaction of the parts but only in their interaction with an outside
world both of the individual parts and the whole.” (Hayek, 1967, pp. 70-71.)
SYSTEMS THEORY, KNOWLEDGE, AND THE SOCIAL SCIENCES 585
directly to read in a logic of operation for the whole from some other source;
and the logic of operation nearest at hand is that of the individual human.
[Hayek, 1979, p. 101; Langlois, 1981, chap. 3.] The notion that, for example,
a society permitting economic self-interest is therefore a greedy society is a
bit of naive individualism characteristic of holists far more than of meth¬
odological individualists.
In order to make a plausible case for holism, we have to argue that less
information is somehow better than more. This is no easy task. We might for
instance invoke the Gestalt and assert that attention to the parts destroys
our understanding of the whole. But this is to confuse perception and under¬
standing. And, as my analogy with literary criticism was meant to suggest,
the sophisticated methodological individualist has no compunction against
stepping back to survey the Gestalt—so long as the (epistemologically more
accessible) parts are also carefully analyzed. Historically, holists have taken
a rather different (if not ultimately unrelated) line of attack. Hegel and his
followers argued that the parts could not be studied in isolation because the
parts acquire their very nature from their relation to the whole, which nature
is necessarily altered if the parts are considered apart from that whole. But if
taken at all seriously, this argument leads to logical absurdity, and it falls
quickly apart when translated from the quasi-mysticism of essentialist
rhetoric into a modern nominalist vocabulary. [D. C. Phillips, 1976, pp. 5-
20.]
Proponents of general systems theory have unknowingly reinvented and
invoked this Hegelian formulation in discussing the holism of systems
theory. [Ibid., esp. pp. 45-50.] But most systems theorists—both mathema¬
ticians and practitioners alike—conceive of systems theory in a way antago¬
nistic to this Hegelian view. They are certainly inclined to recite holist cant
about “phenomena that depend on the conditions of the environment in
which they exist, interact with this environment, and thus cannot be prop¬
erly studied in isolation.” (Mortazavian in this volume.) But all they mean
by this is that the behavior of the whole cannot be understood without
knowledge of the relations among the parts. The parts are conceived of as
logically distinct elements of a mathematical set; those elements exist and
are fully defined independently of any relations that might be specified. A
system is just the set of parts plus a set of relations among the parts.4 This is
a formulation that would trouble a serious holist far more than it would a
methodological individualist.
The philosopher Mario Bunge has recently transformed this set-theoretic
definition of a system into a methodological position—systemism—that
“combines the desirable features of individualism and holism.” (Bunge,
1979a, p. 13.) To Bunge, a society (which he finds it necessary to call a) may
be represented as the ordered pair (5, R), where S is the set of individuals in
the society, and R is the set of relations among them. That systemism does in
many ways combine the best of both worlds is, I believe, an entirely unob¬
jectionable assertion. What is not true is that systemism somehow repre¬
sents a new methodological alternative: The basic ideas of what Bunge calls
systemism are essentially identical to what sophisticated methodological
individualists have believed all along.5
It is important to notice that while Bunge’s definition of society qua
system is consistent with the general mathematical definition of a system
[Mesarovic, 1964a, p. 7], not all applications of systems theory are in accor¬
dance with the tenets of systemism understood as intelligent methodological
individualism. The “parts” in Bunge’s formulation are individuals in soci¬
ety, but this is by no means always the case in social science. The elements
in a typical systems model are aggregate variables of one sort or another—
dollars, commodity levels, energy use—that are represented as impinging
directly on one another with no reference whatever to the existence of
human beings. A good many systems models in the social sciences must,
therefore, be classified as instances of naive holism; and while such models
may often prove interesting and illuminating, they cannot—as is often
claimed—serve as a general foundation for economics and other social sci¬
ences. This is a point to which I will return.
Perhaps I should apologize for the length of this digression. But, in a
sense, perhaps this has not been a digression at all. After all, philosophical
issues lie nearer to the surface in systems theory than they do in most
intellectual endeavors, even if they are not for that reason more often per¬
ceived. More importantly, though, there is a sense in which it is methodolog¬
ical individualism that lies behind the systems-referential view of knowledge
and information that I now wish to present.
5 Bunge does admit that Hayek and other methodological individualists recognize “the reality of
social relations” (p. 17), but he continues to paint them as naive individualists. “The individ¬
ualist might not wish to dispute the systemist’s thesis,” says Bunge, “but, if he is consistent, he
must insist that the structure of R is somehow ‘contained’ in, or deducible from, the properties
of the individual members of society” (p. 19). But, as we saw (cf. again footnote 3), Hayek for
one does not so insist, and I cannot see why he is therefore inconsistent in any way. Bunge is
simply mistaken in his characterization of the individualist position. [Bunge, 1979a.]
SYSTEMS THEORY, KNOWLEDGE, AND THE SOCIAL SCIENCES 587
6This oil-flow model is not without implications. More than one writer has suggested that
information policy be predicated on the inevitable dependency of future society on this "info-
substance” just as energy policy is supposed to deal with our dependence on oil. As a conse¬
quence, we should worry about the availability of “info-fluid" to disadvantaged groups like
Chicanos much as we recently used to fret about the availability of heating oil to poor New
Englanders.
7This is precisely the theme pursued by the British physicist Donald MacKay. [MacKay, 1969.]
The next section of this paper will draw heavily on his ideas.
588 RICHARD N. LANCLOIS
cisionmaker knows the weather—if he knows how much rainfall there will
be—he can easily determine the acreage appropriate for maximum profit. If
he is uncertain about the level of rainfall, he can articulate a probability
distribution for that variable, which allows him to select the acreage that
maximizes expected profit. But the decisionmaker who can anticipate the
level of rainfall perfectly will very likely elect to plant an amount of acreage
different from that selected by the farmer who is uncertain about the
weather; more to the point, the better informed decisionmaker will realize a
higher profit than his ill-informed counterpart.
If the decisionmaker who is uncertain can obtain some kind of informa¬
tion about the weather—divine revelation, perhaps, or the weather bureau’s
probability distribution for rainfall—then he can improve his acreage deci¬
sion. But if this information is costly, the decision whether to acquire it can
be represented as yet another decision-system of the same sort as the origi¬
nal acreage decision. In general, our profit-maximizing farmer should wish
to acquire the information if its expected value exceeds its cost. [See, for
example, Howard, 1966.]
This expected value of information is the interesting quantity. In order to
examine it more closely, let us simplify our example even further. Suppose
there are only two possibilities, heavy rain and light rain. The farmer must
optimize his acreage planted in light of whatever information he might have
about which of these two possibilities will occur; for example, he might
consider the amount of rain experienced in past years or various freely
available predictions by the Weather Bureau or the Department of Agricul¬
ture. On the basis of this imperfect information, the farmer estimates a
probability of, say, 0.7 for heavy rain and 0.3 for light rain. Now, suppose
that a fully reliable clairvoyant stands willing (for a stiff fee) to disclose
which of the two alternatives nature actually has in store. How much is this
revelation worth to our profit-maximizing decisionmaker? We have to con¬
sider how the new information would affect the farmer’s acreage choice and,
thus, his profits. If the clairvoyant says “heavy rain,” he can optimize the
acreage in a way that increases profits over the expected level. If the clair¬
voyant sends the message “light rain,” he can also optimize acreage—in a
different direction—to increase profit over the expected profit. In deciding
whether to buy the clairvoyant’s information, the farmer must decide if the
increase in profit will be enough to justify the clairvoyant’s fee—and he must
decide before he knows which message the clairvoyant will send. So, the
decision whether or not to buy the information must be based on the farm¬
er’s prior probability assessment of what the clairvoyant will say. Since the
clairvoyant is merely revealing what nature will do, the farmer’s weather
forecast or other prior information is exactly as relevant to predicting the
clairvoyant’s message as to predicting nature’s response—since the former
is nothing but the latter moved forward in time.
Thus, the farmer must use his original probability assessment (0.7 chance
of heavy rain; 0.3 chance light rain) in deciding whether to buy the informa¬
tion. And the expected value of perfect information is thus the sum of two
590 RICHARD N. LANGLOIS
magnitudes: the first is the increase in profits from adjusting to heavy rain if
there will, in fact, be heavy rain, multiplied by 0.7, the probability that there
will be (and, therefore, that the clairvoyant will say) heavy rain; the second
is the increase in profits from optimizing for light rain given that there will, in
fact, be light rain, times the probability that there will be (and, therefore,
that the clairvoyant will say) light rain.9 In other words, the expected value
of perfect information (EVPI) is the sum of the value to the system of a set of
possible messages weighted by the probability of occurrence of each mes¬
sage.
By contrast, the information measure of communications theory is oblivi¬
ous to the value a message holds for the system that receives it: In analogy
with the entropy measure of thermodynamics, the information content of the
message heavy rain is, to the communications theorist, proportional only to
the logarithm of the probability of that message being received.10 Although
the value measure and the entropy measure can move in the same direction,
there is no general reason why this should be so.11
9The length of this sentence suggests that there are times when mathematics has its expository
(or at least space-saving) advantages. Let El(a|w) be the (realized) profit from planting a acres
under weather conditions w. The farmer’s original problem is
where H indicates heavy rain, L light rain, and p is the farmer’s prior probability assessment on
heavy rain. (The probability of light rain has to be (1 - p) since there are only two possibilities.)
Suppose that a* is the value of a—the number of acres—that maximizes the quantity in
brackets. Then,
where p is the farmer’s assessed probability that the clairvoyant will call for heavy rain (which is
necessarily identical to his original assessment of the probability of heavy rain), a*H is the
optimal choice of acreage under conditions of heavy rain, and a*L is the optimal acreage under
light rain conditions.
10In the notation of footnote 9, the (nonsemantic) information content of the message heavy
rain is proportional to -log2p. The information content or entropy of the light-rain/heavy-rain
information system is the information content of each possible message weighted by its proba¬
bility,
“In our now familiar example, it happens that the two measures are closely related: Informa¬
tion content is high when value is high and vice versa. (The reason is that a high-entropy [low
“info-content”] message is one with a high probability. If a farmer anticipated heavy rain with a
high probability, then a* is likely to be already near a*H, and [II(a^|//) - II(a*|//)] is low.
Similarly, the more unexpected message would have both a higher information content and a
higher value, since a* would be less close to a*. I believe this monotonicity can be shown
rigorously to hold if the profit function is concave.) But such a connection is merely fortuitous.
Consider an industrial research laboratory, for example, in which one experiment has a very
high “info-content” (e.g., the experiment has a 50-50 chance of resulting in the message suc¬
cess) but low profit implications for the company, while a second experiment has a low “info-
content” (e.g., a 90 per cent chance of success) but big profit implications. The research
manager who used entropy as a decision-making criterion would be sorely misguided.
SYSTEMS THEORY, KNOWLEDGE, AND THE SOCIAL SCIENCES 591
It isn’t until we consider the range of other states of readiness, that might have
been considered but weren’t, that the notion of meaning comes into its own. A
12By a cybernetic system in a broad sense, I mean any system that is concerned with informa¬
tion and control. In a narrower sense, cybernetics is concerned specifically with feedback
systems, where a monitoring signal is sent from output back to input in order to control that
output and bring the system into an equilibrium condition called homeostasis. The locus
classicus here is Norbert Wiener’s Cybernetics. [Wiener, 1948 and 1961.]
13 “Depuis longtemps, le pragmatisme et le behaviourisme ont appris aux psychologues a mettre
Taccent sur Taction plutot que sur la conscience. La cybemetique adopte rigoureusement ce
point de vue: le sens, la conscience dans l’information, n’a rien d’essentiel; ou plus exactement,
le sens d’une information n’est rien d’autre que Tensemble des actions qu elle declenche et
controle.” (Ruyer, 1954.)
14Here, I think, we need to make the distinction between the (1) meaning, (2) meaningfulness,
and (3) value of a message. A message to the decisionmaker reading keep doing what you’re
already doing is fully meaningful even though it does not entail a different action or necessarily
result in a level of goal achievement higher than would have occurred in its absence. (Of course,
a strict behaviorist would be unable to distinguish a meaningless message from a meaningful
message to maintain the status quo, but this is not a problem in a goal-directed model.)
592 RICHARD N. LANCLOIS
lsAnd here, in an important sense, is where methodological individualism comes back into the
picture.
"’This is true of Wittgenstein in the Tractcitus logico-philosophicus and also of the later logical
empiricists. [Wittgenstein, 1922; Carnap, 1950; Bar-Hillel and Carnap, 19536; Bar-Hillel, 1964.]
I should also note that it was this attempt to eliminate the personal and (contra Mortazavian in
this volume) not the use of subjective probability that was the ultimate problem with the
positivist approach. It was a problem of too little subjectivism, not too much.
SYSTEMS THEORY, KNOWLEDGE, AND THE SOCIAL SCIENCES 593
l7We can see this clearly by comparing the equations in footnotes 9 and 10. For the value-of-
information measure, we had
where Alt is shorthand for the expression in brackets in footnote 9. For the entropy measure,
we had
In the first case, the probabilities weight terms that measure the effect of a message on the
system; in the second case, the probabilities weight terms referring only to the selection of a
message from the set of messages.
594 RICHARD N. LANGLOIS
core memory, for example, is changed when data are stored. Furthermore,
the memory bank is not the only locus of memory in the system. In an
important sense, the entire organization of the system—hardware plus soft¬
ware, mind and body—contains functional knowledge to guide behavior. I
do not think this is a new or particularly controversial way of putting things.
At the risk of a charge of idiosyncrasy, let me go beyond this to suggest a
distinction between structural information and parametric information. The
former is information that operates on—that changes the basic structure of
the system; the latter is information that operates on parameters of the
system—on elements that adjust or calibrate the workings of the system
within the dictates of, but without altering, its underlying structure.
I apologize if this all sounds a little vague. But the nature of form and
structure is a problem that has animated philosophy since the Presocratics,
and I make no pretense of trying to solve it. At an intuitive level, the notion
of a system’s structure is fairly clear: A system has various fixed (or rela¬
tively slow changing) attributes that define its form, that set the scope,
ground rules, and boundary conditions for the system’s more variable as¬
pects. (This may have a physical correlative in hardware as distinct from
software, but it need not; even within the software of a computer program,
the mental representations of a human mind, or indeed any system in the
abstract, we can speak of an underlying structure.) If we restrict ourselves to
the mathematical realm, we can make the notions of structural and
parametric information more precise, if in the end perhaps no less
metaphoric.18
The agricultural decision-system again provides a clear example. Here,
the model’s structure lies in the form of the profit function; that function
specifies a parameter, weather conditions, that can be altered by reception
of an appropriate message. Our decisionmaker is able to obtain only
parametric information and thus to gain only parametric knowledge from a
signal. As the problem is formulated, no signal can change his profit function
or any of the basic givens of the situation he faces.
In the modern mathematical “economics of information [Hirshleifer and
Riley, 1979], the focus is exclusively on parametric information of this sort.
For a long time, the mainstream of economics had concentrated on “perfect-
information” models, where the decisionmakers were portrayed as having
full knowledge of all aspects of the decision-situations they faced. In newer
18Mesarovid offers a definition of systems structure that, as best I understand it, is consistent
with what I have in mind. The organization of a system consists in the systems relation that
maps one set of parts into another. This relation R “can be considered as defined by an abstract
relation and the specific values of the unspecified constituents, the so-called relational con¬
stituents; R = {T, £}, where T = systems structure, £ = set of relational constituents.”
(Mesarovid, 1964a, p. 10.) As Mesarovid further suggests (in Equation 9), these relational
constituents can take the form of parameters of the system. Thus, my distinction between
structural and parametric seems in accord with his view. Also see Mortazavian’s paper in this
volume.
SYSTEMS THEORY, KNOWLEDGE, AND THE SOCIAL SCIENCES 595
It is fair to conclude, I think, that systems theory and the theory of knowl¬
edge and information (broadly defined) must ultimately be related in a funda¬
mental way. Both are concerned with form and organization. And it is little
wonder that communications theory, with its entropylike measure of infor¬
mation content, has held a singular fascination for systems theorists. An
organized entity is a nonrandom entity, one whose organization is unex¬
pected in some sense, and it is unexpectedness—negative entropy—that
communications theory measures.
Although, as we saw, the derivation of a scalar measure of organization is
possible only in the very special case with which communications theory is
concerned, it nonetheless remains that the knowledge content of a system
is closely bound up with that system’s organization—with its structure.
We have already seen two logics of organization identified by systems
theorists: causal systems and goal-directed systems. I would now like to
suggest that these two are not the whole story.
A causal system is an instance of what we might call a “mechanical”
structure. The movements of the parts are causally related to one another
within the dictates of a fixed structure. The centuries-old example is the
mechanical clockwork. Each gear moves by virtue of the force impressed on
it by a previous gear; and each carries out its function within the pattern
ordained by the designer. Once cut loose from its creator, the mechanical
system cannot increase its level of organization. Indeed, any change in
structure (other than those effected by the ministrations of the designer)
must lower the level of system organization, thus increasing entropy in some
sense.
Information and control are closely related concepts in systems theory. In
a strictly causal system, the only way to change behavior is by reprogram¬
ming the system; often, this can be accomplished by adjusting various con¬
trol variables to modify the system’s structure—much as we manipulate the
steering wheel and foot pedals to alter the behavior of an automobile. This is
596 RICHARD N. LANCLOIS
191 should also mention here the branch of systems theory called optimal-control theory. The
optimal-control theorist seeks the best way to control a system in order to achieve some goal.
For example, we might wish to calculate that trajectory of a rocket between two points that
minimizes flight time, fuel consumption, or some other objective. (And, as a matter of fact, it
was precisely such aerospace problems that formed much of the early subject matter of this
theory.) Mathematically, the optimal path is found using the calculus of variations (and some¬
thing called the Pontryagin Maximum Principle) or the (closely related) technique of dynamic
programming. Once found, this optimal trajectory can be imposed directly by open-loop control
or implemented through a feedback system to create goal direction and closed-loop control.
[See generally Bryson and Ho, 1968.] Optimal-control theory has found its way into economics,
especially in economic-growth models (not unrelated to the simple stabilization policy models
already mentioned) [Burmeister and Dobell, 1970, chapter 11] and in the theory of resource
extraction [Sweeney, 1977]. See also Kamien and Schwartz [1981] and Aoki [1976].
20Human biology was, in fact, an early inspiration for cybernetics. [Cannon, 1939.]
SYSTEMS THEORY, KNOWLEDGE, AND THE SOCIAL SCIENCES 597
innovation is a potential victory for the forces of entropy and disorder over
the forces of homeostasis. [Langlois, 1981, chap. 2, esp. pp. 73-82.] And it is
thus probably no surprise that systems theorists [Mesarovic and Pestel,
1974] and cybernetically oriented biologists [Hardin, 1977] are prominent
among the new Malthusians who see dangerous social and ecological imbal¬
ances in the world’s future.
In economics, there has been some effort afoot, especially by so-called
post-Keynesians, to close the loop on the macroaggregate models of eco¬
nomic stabilization. These economists, as one observer correctly notes,
“. . . might wish to replace the Newtonian-clockwork model by something
they call a ‘cybernetic’ model, which may be an improvement (if it could
ever be devised), but a shift from mechanical statics to sophisticated
mechanical dynamics is no radical conceptual revolution.” (Kristol, 1981, p.
212.)
What is the alternative to the mechanical causal and cybernetic models?
The answer is best found by leaving the level of aggregate dynamics and
returning to consideration of the parts—the human agents. A human being—
at least in part or at times—is a goal-directed system. This is the basis of the
ideal type—homo economicus—that underlies much of mainstream eco¬
nomic modeling. Yet everyone this side of B. F. Skinner recognizes that the
flesh-and-blood human being does not operate with the stimulus/response
compulsiveness of a simple goal-directed cybernetic system. There is some¬
thing more, or perhaps different, at work.
One major school of thought holds, at least implicitly, that differences
between the human mind and a mechanical cybernetic system are ones of
degree rather than of kind. The mind is an immensely complex thing, and if
we could only construct a cybernetic system—that is to say, a digital com¬
puter, the apotheosis of the mechanical cybernetic system—with sufficient
complexity, we could largely replicate much of what the mind can do. Ad¬
herents to this view take heart from the old saying attributed to Marx that
quantitative change allowed to go on long enough inevitably becomes qual¬
itative change.
There are dissenters from this viewpoint, of course. Most notable among
these is Hubert Dreyfus, whose analysis continues to stir controversy in the
field of artificial intelligence. [Dreyfus, 1979.] In a significant sense, Drey¬
fus’s thesis rests on the system-referential view of knowledge already ar¬
ticulated. Human knowledge, he argues, is very much tied to the biologically
and culturally evolved structure of the human organism. What is meaningful
to a human is meaningful only in reference to that structure, and meaning
cannot be reduced to the system of explicit, extrapersonal, context-free
statements that a computer, by virtue of its own structure, must employ. It is
for this reason, Dreyfus argues, that no artificial intelligence program has yet
been (or could be) written in which the important functions of discriminating
meaning and significance are not preprogrammed by the human designer.
However this be resolved, it is clear that the human mind operates differ-
598 RICHARD N. LANGLOIS
21 Ludwig von Bertalanffy’s definition of this term seems a bit different and comes nearer to the
distinction I am looking for. He describes a self-organizing system as one capable of “evolving
from a less to a more differentiated state.” (Bertalanffy, 1968, p. 68.)
22The reason for this, I believe, is that Mesarovic is willing to classify systems as self-organizing
if they change their structure within the dictates of a fixed higher level structure; that is, if they
“change their structure by using a relation from the set fl^.“ (Mesarovib, 1964«, p. 11.)
23Morphogenetic man has not been entirely neglected in economics. Under the title of entrepre-
SYSTEMS THEORY, KNOWLEDGE, AND THE SOCIAL SCIENCES 599
If economic agents are cybernetic reactors, they can adjust the economic
mechanism in light of changed circumstances, but they cannot thereby in¬
crease the organization (decrease the entropy) of that mechanism. And any
imperfection in the agent’s knowledge (that is, any lack of correspondence
between the economic problem perceived by the agents and the “true”
economic problem) can lead to a bad equilibrium, a market failure. But if
economic agents can alter the problem they face, if they can bring new
structural knowledge into the system, then elaboration of, and increased
differentiation in, the economic system becomes possible. Far from causing
disorder or chaos, apparent departures from homeostatic equilibrium can
actually result in an increase in system organization and a decrease in en¬
tropy. This is the phenomenon of “spontaneous order” [Hayek, 1967, p. 77]
or, in technical jargon, “deviation-amplifying mutual causal effects.”
[Maruyama, 1963.]24
Recognizing the existence and importance of morphogenetic processes
has a number of far-reaching implications for both systems theory and eco¬
nomics, but this is not the place to explore them, I am afraid. Instead, let me
close with some brief observations on relations between these two disci¬
plines.
Systems theorists of my acquaintance are sometimes inclined to the opin¬
ion that economics lacks an adequate systems perspective and its scientific
development would be rapidly enhanced by a complete subsumption of that
discipline into systems theory. This viewpoint is not without its ironies. In
an important sense, it was economists who were the first systems theorists.
Moreover, the founders of economics were concerned with a fully mor¬
phogenetic version of systems theory. [Hayek, 1967.] Adam Smith’s concep¬
tion of economic growth based on the increasing division of labor is very
much a theory about the evolution of economic structure from a less to a
more differentiated state. [A. Smith, 1776 and 1936.] And, indeed, it is now
widely recognized that the theory of evolution, known to us through the
biological theory of Darwin, was articulated at least a century earlier in the
social sciences.25
Of course, modern systems theory is also concerned with questions of
morphogenesis, but mathematical systems theory—in the guise of non¬
equilibrium thermodynamics [Prigogine, 1971] and topology [Thom,
neur, he has been studied by Joseph Schumpeter and more recently, Israel Kirzner, among
others. [Schumpeter, 1934; Kirzner, 1973.]
24 Another name for this class of phenomena is autopoiesis, an area that is apparently attracting
increasing interest among systems theorists. For a bibliographic introduction, see Zeleny
[1981].
25David Hume, who numbered economics and social thought in general among his philosophical
interests, adumbrated a genuinely Darwinian view of evolution in his Dialogues Concerning
Natural Religion, first published in 1779 but written in 1759, a hundred years before The Origin
of Species. [Hume, 1961, especially p. 478.]
600 RICHARD N. LANGLOIS
26This work by Thom, which presents the notion of catastrophe theory, has attracted a large
following in a number of disciplines and remains something of a cause celebre in mathematical
systems theory. It promises mathematical interpretations of some of the most ancient and
vexing problems of philosophy—the nature and origin of form and structure and the text
alternates enticing intimations of the profound with an utterly impenetrable formalism, thereby
applying a time-tested formula for attracting a cult following.
27On the similarities between physiocrats and certain modern systems models of the economy,
see Almarin Phillips [1955].
SIGNALS, MEANING, AND
CONTROL IN SOCIAL SYSTEMS
Walter Buckley
There are three main topics in Langlois’s paper on which I would like to
comment. One is his philosophical critique of systems theory involving
“emergentism.” Second is his discussion of information and meaning in
relation to social science. Last is the question of control theory, again as it
may apply to social science.
I shall not dwell long on the first topic since, in my view, Langlois has
erected a straw man in arguing that systems theory, in its concern for the
whole and its emergent properties, tends to ignore the components. None of
the founding theorists or important contributors have held such a view. The
mistake perhaps derives from the emphasis given to emergent properties and
behaviors of the whole in the attempt to offset the strong reductionist ten¬
dencies in most of the sciences. This issue has been well reviewed by Anatol
Rapoport in his contrast of the traditional analytical methodology of physics
with the glaring need for synthetic methods in the biological and social
sciences. [Rapoport, 1968«.]
Langlois’s characterization of intelligent or sophisticated methodological
individualism wipes out any significant differences with emergentism and
does away with important issues that still need to be thrashed out in princi¬
ple. In practice, however, since we cannot begin to explain or predict the
properties or behaviors of complex dynamic wholes from a knowledge of the
parts, it seems wise to study both and try to understand how each interacts
with and influences the other.
In discussing communication theory, it is, of course, important to distin¬
guish and define signals, information, and meaning and distinguish as well
their relevance to simple systems as compared to complex systems that
decide or choose different behaviors. Soon after Shannon published his
theory of information, it became widely recognized that it was actually a
theory of signals, and he himself recognized that it did not deal with informa-
601
602 WALTER BUCKLEY
See Wassily Leontief s recent critique of academic economics and its nonempirical founda¬
tion. [Leontief, 1982.]
SIGNALS, MEANING, AND CONTROL IN SOCIAL SYSTEMS 605
not by itself provide a theory for any particular substantive area of study. To
be useful, it must be integrated into the conceptual and empirical materials
of a discipline. There are too few who have mastered it sufficiently well to
apply it knowledgeably to their social-scientific field of study, especially
empirical study. Most systems researchers are not social scientists and vice
versa. Consequently, many attempted applications of systems theory to
aspects of society are either poor social science or poor systems research. It
is asking a lot to expect someone to master another field of study such as
systems research, especially when the latter itself draws from a number of
disciplines. But some of its important concepts are being quietly assimilated
by various social sciences, though substantial progress may have to wait for
the development of institutes and programs designed to integrate systems
research and substantive studies.
It is my belief, however, that the promise of systems research for the
social sciences is still great and has much more to offer in the areas of
communication and social control than we have yet seen. The two need only
engage in more intensive and empirically relevant interaction.
TOWARD A SYSTEM-BASED
UNIFIED SOCIAL SCIENCE
Robert D. Beam
We often hear that "one man can no longer cover a broad enough
field" and that "there is too much narrow specialization" . ... We
need a simpler, more unified approach to scientific problems, we
need men who practice science—not a particular science—in a
word, we need scientific generalists.
Richard Langlois’s paper reflects the opinion that systems theory has failed
to bring about interdisciplinary synthesis of, and integrated education in, the
social sciences. I would like to present a quite different relation between
systems theory and the social sciences. In contrast to Langlois’s position, I
would like to suggest a means by which systems theory can “unify ... the
methodology of all sciences, . . . endow the social sciences with that natural-
science-like rigor,” and combat the “fragmentation and specialization” of
social-science knowledge. (Langlois in this volume.) It is my opinion that
systems theory has much to offer at the basic and most general level of social
science, because, and again in contrast to Langlois, I feel it has succeeded in
bringing “all disparate parts of inquiry under the sway of its organizing
force.” (Ibid.) Systems theory has provided a framework for social theory, a
framework that allows for unification of the behavioral and social sciences
on an analytic-deductive basis. Such a unified social science is built on
systems concepts from the ground up, and the chief objective of my com¬
ment is to provide a brief outline of this discipline as it relates to a selected
few of Langlois’s remarks.
I would first like to articulate a model that makes a clear distinction
between the servomechanism responding to environmental stimuli in a pre¬
programmed fashion and the human system as a complex controlled system
capable of effecting conscious learned responses to environmental stimuli.
This distinction is, I believe, one that Langlois also considers to be required.
A complex controlled human system seems to me to be analogous to what he
terms a metacybernetic system. This is a model that seeks to accomplish for
607
608 ROBERT D. BEAM
social science what the model of “economic man” accomplishes for eco¬
nomic science. In contrast to the cybernetic-reactor model of economic
man, this model enables the agent to alter the environment or his or her
relation to it.
Since all behavior of analytic relevance to social science is learned behav¬
ior, this model focuses only on behavior that is directed from the nervous
system toward some goal. The present model also assumes that all informa¬
tion that can be used to direct behavior is stored in the form of concepts
(images) to be retrieved from storage through perceptions, memory, or com¬
munications.
The simple cybernetic model includes three logically irreducible ingre¬
dients for purposive behavior selection: a detector that acquires and pro¬
cesses information about the system’s environment; a selector that chooses
responses on the basis of the system’s goal-state; and an effector that effec¬
tuates the overt behavioral response. The nature of control is the system’s
internal ability to detect deviance from a preferred state and correct its own
behavior so as to restore, or move back toward, the “preferred” position.
All purposive behavior requires negative feedback for both servomecha¬
nisms and organisms, including the human system.
But unlike other controlled systems, humans possess conscious aware¬
ness of, or information about, the system states of their detector, selector,
and effector (DSE) subsystems. The more complex model of the human
system contains sub-DSE controls for each of the main-level DSEs.
Mechanical man-made servomechanisms, even computers, do not possess
these types of sub-DSE controls.1 The complex cybernetic model presented
here provides a conceptually simple explanation of consciousness, even
though it is not yet firmly established where in the brain consciousness
resides or how it functions. Although the human biological system is respon¬
sible for effectuating overt behavior, it is the method of behavior selection
that is of interest to social science. Thus, it is the control system of humans,
as opposed to the biological system, that formulates and guides the execu¬
tion of behavior relevant to psychological and social analysis.
The terms detector, selector, and effector have strictly functional defini¬
tions. The detector encompasses all sensory, cognitive, and intuitive pro¬
cesses by which humans form, identify, and modify both real and imaginary
concepts about themselves and their environment. The selector spans the
entire range of emotions, pleasures, pains, fears, loves, desires, and so
forth—all things to which positive and negative valences can be attached, a
valence being a property of the nervous system that produces pleasant or
unpleasant feelings that reinforce approach responses or avoidance re¬
sponses, respectively. The effector includes all neural connections that di-
1 For a more complete presentation of the human system as a complex controlled system, see
Alfred Kuhn [1974, pp. 60-102; 1975, pp, 49-73], The contents of this paper draw heavily on the
model of unified social science, as developed in these two volumes.
TOWARD A SYSTEM-BASED UNIFIED SOCIAL SCIENCE 609
In this model, the detector deals with cognition and perception, which are
essentially the processes of learning and using patterns. Cognition, or pat¬
tern learning, involves the processes by which pattern concepts are orga¬
nized and stored in the detector. Perception involves the use of incoming
sensations as cues to identify (match) events in the environment with con¬
cept images (referents) of those events retrieved from storage. More
specifically, signal inputs from the environment activate sensory receptors
that modulate sensory nerves. This is the detection stage in which informa¬
tion received is without meaning. Information acquires meaning only after
the detector codes or organizes sensory and linguistic inputs into concepts
or patterns of concepts by categorizing the inputs by their similarities and
differences. The detector can organize information into its own concepts or
it can adopt the organized concepts of others through communication. The
detector forms new concepts by abstraction (inductively forms new “emer¬
gent” concepts through the essentially morphogenic process of random
variation and selective retention or intersection (acquires new concepts
by deduction and discrimination, usually in the form of semantic com¬
munications). Knowledge in the detector is thus organized information, and
this, I take it, is what Langlois means by his statement that knowledge and
organization are identical.
If cognition in the detector is the process of coding and storing informa¬
tion in conceptual bins, perception is the process of decoding or retrieving
coded information from storage. Perception is the process by which the
detector matches primary patterns of incoming sensations with secondary
patterns of previously stored images (knowledge) to infer information about
presently experienced reality that is not contained in either primary or sec¬
ondary form alone. Perception thus involves the receipt and use of primary
information as cues to identify or activate secondary information in the form
of previously stored concept-images. Matching uncoded sensory inputs with
previously coded (learned) and stored concepts is known as pattern recogni¬
tion. A signal input does not become a stimulus in the detector until it has
been decoded or identified, and strictly speaking, it is not the signal itself,
but the identification that constitutes the stimulus. A meaningful signal input
is one whose cues are sufficient to activate a stored channel of concept
images in the detector, or in Mac Kay’s terminology, to select from the range
610 ROBERT D. BEAM
For the individual human, all decision processes are confined to the detector
and the selector. Effector processes are confined to the muscular execution
of previously selected behavior. Pattern learning in the effector involves
learning new performance skills, while pattern using involves their exercise.
Less emphasis is placed on the effector in understanding directed behavior
in humans, since social science is concerned more with the what and why of
behavior selection than the how.
A person’s behavior is guided toward his or her goals by decisions that
reflect the interrelated detector, selector, and effector states of his or her
control system. Decision theory deals with response selection in humans
under conditions of complexity, particularly with respect to the conscious
content of all DSE subsystems. The adaptive behavior process can be
viewed in this model as involving two stages, a performance stage and a
feedback stage. During the performance stage, the detector identifies an
opportunity function of perceived alternative responses consonant with the
TOWARD A SYSTEM-BASED UNIFIED SOCIAL SCIENCE 611
Explicans
Premise 1 (detector) if: our decisionmaker perceives that planting acreage
Ax will yield a profit of $100,000.00 and that planting
acreage A2 will yield a profit of $75,000.00,
Premise 2 (selector) and if: his dominant motive is to maximize profits,
Explicandum
Conclusion (effector) then: he will plant acreage Ax.
All disciplines in the social sciences deal with types of social systems, yet
their specialized concepts bear little resemblance to one another. There
exist, however, some basic social mechanics that underlie all social science
in the same sense that principles of the lever, inclined plane, and pulley
underlie all mechanical contrivances. [Alfred Kuhn and Beam, 1982.] The
intrasystem/intersystem axes of controlled systems shown in Figure 1 reveal
basic system-based concepts of social interaction that underlie not only the
core disciplines of economics, political science, and sociology, but also such
related areas as small group theory, social psychology, management theory,
and organization theory. The rectangles in Figure 1 show two goal-oriented
systems, to wit, two individuals or two formal (controlled) organizations of
individuals. Within each system are listed the three logically irreducible
ingredients for purposive behavior: detector, selector, and effector. Behav¬
ior selection is separated into a decision stage and an effectuation stage with
feedback. An intrasystem view consists of analysis of the behavior of a given
system as a function of the DSE states of its control system. Since social
science deals with interactions among two or more goal-oriented systems, an
intersystem view is used to analyze interactions of their detector subsystems
(communication), their selector subsystems (transaction), and their effector
subsystems (organization).
Communication is the process of transferring patterns between systems.
It both reflects and affects their detector states. Transaction transfers things
of value between systems and reflects and affects their selector states. Or¬
ganization carries out joint behavior by two or more systems and reflects and
affects their effector states. Typically, systems theorists have analyzed link¬
ages (social interactions) among humans with the analytic tools in informa¬
tion and communication theory. [Cadwallader, 1959; Haberstroh, I960.]
Matter/energy interactions—for example, sexual union or physical com¬
bat—are typically categorized in a different conceptual bin. [Buckley, 1967,
p. 48.] The intersystem view of Figure 1 separates interactions analytically
into the science of pattern transfers and the science of value transfers across
controlled systems. Although considered the sine qua non of interaction,
communications are viewed in this model as the facilitators of transactions
that reflect the values people attach to their matter/energy transfers, and not
the matter/energy transfers themselves. Thus, communications and transac-
TOWARD A SYSTEM-BASED UNIFIED SOCIAL SCIENCE 613
Party A Party B
tions are the interactions between humans that parallel the information and
matter/energy interactions between living systems. I am not at all certain
how Langlois would respond to the contention that a tight and coherent
model of social interaction should include both the science of communica¬
tion and the science of transaction,2 each of which can be briefly outlined as
follows. [Alfred Kuhn, 1974, chaps. 7-11.]
Communications
2 A more extensive coverage of the propositions relating to these basic sciences can be found in
Alfred Kuhn [1974], chaps. 7-11.
614 ROBERT D. BEAM
concerned with the semantic level of communication, but with the amount of
information measured in binary digits that can be accurately transmitted
through the channel.
By contrast, the communication model in Figure 1 assumes that the chan¬
nel is capable of accurately transmitting information—signals—and is inter¬
ested instead in the accuracy with which (1) the coded pattern to be trans¬
ferred is clear in the detector of the sender, (2) the message is detected by
the receiver, and (3) the concept-referents that the transmitted signs select in
the receiver’s head correspond to patterns in the sender’s head. Analytic
interest is here directed to the effect of the communication on the detector
states of the interacting parties, and not on the channel capacity of the
transmission-medium or its levels of noise or redundancy.
I believe this view of semantic pattern-transfer is in accord with Mac-
Kay’s framework as Langlois describes it. It is an alternative to the oil-flow
model of information, which, like the systems-referential view, might argue
that information is not like “hardware products to be delivered through
obstacles with minimum damage at least cost.’’ (Nevitt, 1981, p. 595.) The
core of the semantic problem of communication for social science lies in the
proper selection of signs, such that they serve as cues sufficient to enable the
receiver’s detector to decode their meaning. The mathematical theory of
communication, with its emphasis on bit-counts, noise, entropy, redun¬
dancy, and so forth, has much to contribute to binary computers, telecom¬
munications, and other areas. But it has little to say about the manner in
which humans match signs with concept images or recombine these images
through sign communications to form new images. Although bit-count analy¬
sis can apply to information transmissions below the social level, it does not
determine how much meaning is transferred by a given sign. The meaning of
the sign depends on the concept-referent of that sign as it exists in the head
of the sender at the moment of transmission and within the context in which
the sign is used. The communication model in Figure 1 is concerned with the
transfer of patterns across detectors of separate individuals, and not with
transmission of information in bits. Hence, a discussion of bit-counts when
referring to pattern transfers is not directly relevant to the analysis.
Transactions
Organizations
Whenever a continuing and reasonably stable pattern emerges from the joint
effects of the separately decided behaviors of two or more parties, the rela¬
tion they form is called an organization. An organization is a social system of
interacting subsystems of human components. It is an emergent social phe-
616 ROBERT D. BEAM
nomenon in that it is a higher level system than its component members. The
whole of an organization is more than the sum of its parts, where sum is
taken to mean not their numerical addition, but their unorganized aggrega¬
tion. [Buckley, 1967, p. 48.]
Just as an analytic need arose to separate interactions into communica¬
tions and transactions, so there is also an analytic need to separate an
organization by its structure and its process. Its structure is described in
terms of its subsystems and their interrelated roles, a role being a set of DSE
states and actions of a subsystem, including its specified communications
and transactions with other subsystems. A role is thus a pattern-set of expec¬
tations about the behavior of the individual who occupies that role, and the
structure of roles is analogous to the blueprint of the organization. The
organization process describes the interactions of the role occupants them¬
selves, particularly the way in which their actual communications and trans¬
actions constitute the joint effects of the organization as a whole/ [A. Kuhn
and Beam, 1982, chaps. 8-9.]
As social systems, organizations span the analytic gamut from purely
controlled (formal) organizations to purely uncontrolled (informal) organiza¬
tions, with a continuum of mixtures (semiformal organizations) in between.
Formal organizations, such as governments or business enterprises, possess
a controlled subsystem that acts on behalf of the whole system and has the
capacity to communicate with, induce motivations of, and instruct its sub¬
system role-occupants to behave in the way it (the control system) desires.
[Ackoff, 1957-1958.] Formal organizations are thus oriented around the con¬
sciously coordinated behaviors of parts into behavior of the whole. Con¬
scious coordination plus whole-system goals distinguish formal organization
from its conceptual opposite, an informal organization. An informal organi¬
zation such as a pure market economy is an ecological social system that has
no control subsystem, no whole-system goals, and does not behave as a unit.
Each subsystem pursues its own self-oriented goals, and any coordination
that produces a joint effect of their interactions involves morphogenic pro¬
cesses that are wholly unplanned and undirected at the main-system level.
The main identifying criteria of informal organization are absence of con¬
scious coordination and presence of attention to subsystem goals. A semifor¬
mal organization such as the sociocultural system is informal in the sense
that it does not behave as a unit but formal in that selected subsystems that
do behave as units seek to modify their own behavior as well as the behav¬
iors of other subsystems in ways they believe are in the best interests of the
organization as a whole. Thus, to repeat, the main identifying characteristics
of semiformal organization are absence of conscious coordination and pres¬
ence of attention to whole-system goals by subsystems.
3 See the relation between an organization as an acting system and a pattern system, in Alfred
Kuhn and Robert Beam [1982], chaps. 8-9.
TOWARD A SYSTEM-BASED UNIFIED SOCIAL SCIENCE 617
Every discipline in the social sciences is a set of lenses and a set of blinders.
Our academic institutions have been quite successful in providing special¬
ized lenses but have been less successful in removing the blinders. The
intrasystem/intersystem axes of controlled systems in Figure 1 are a tight,
logical way to unify the methodology of all social science disciplines. Re¬
gardless of the specific context in which it occurs, the study of behavior-
selection in humans can focus on the three irreducible components of goal-
oriented behavior: detector, selector, and effector, or DSE. Coordinated
action among humans is possible only by linking together their control sys¬
tems via their detectors (through communication), their selectors (through
transaction), their effectors (through organization), or some combination of
these. Communication, transaction, and organization are the most basic
building blocks of social analysis. When used as modular units, these gen¬
eral-purpose conceptual basics can be assembled into a vast variety of spe¬
cial-purpose configurations that lie scattered throughout the conventional
social science disciplines.
For example, supply-and-demand curves in economics can be derived
from the transaction model by adding parties to each side of an initial trans¬
action. [Alfred Kuhn, 1974; Beam, 1979.] Conflict and cooperation are
configurations of communication and transaction. Transaction has also been
successfully used as a framework for labor negotiations. [E. E. Herman and
A. Kuhn, 1981, chaps. 11-12.] The symbolic interaction process is a
configuration of communication, transaction, and perceptual-detector pro¬
cesses. Authority is a unique transactional interaction within the context of
formal organization. Decision theory is the study of complex DSE processes
under conditions of uncertainty. The theory of government, as modified by
conditions of sovereignty, all-purpose goals, involuntary membership, and
legitimacy, is a special-purpose configuration of the formal organization
model. [Alfred Kuhn, 1975, chap. 13.] Group decision-making is achieved
through communication (persuasive influence), transaction (trade-off or
compromise), or a particular transactional configuration known as dominant
coalition. In each case, the same few basic components are fitted together as
modular units to form special-purpose configurations previously viewed as
the exclusive domain of specialists in each of the social science disciplines.
The purpose of this approach is not to undercut the disciplines, but to
provide them with a common set of system-based analytic underpinnings.
The integration of social science knowledge is not to be achieved by setting
more and more areas of specialized knowledge side by side under some
thematic interdisciplinary umbrella. [Bowler, 1981, p. 214.] Integration is
achieved, rather, by developing more efficient conceptual structures that
allow the human mind to deal with larger amounts of information by means
of a smaller number of general concepts. It is achieved with a structure that
618 ROBERT D. BEAM
Richard H. Day
pactly as much as we know and can find out rigorously about classes of
disparate phenomena. In short, systems thinking is a means for economizing
thought.
No one should appreciate such a point of view more than economists. The
principle is economic. Its application is in no field better exemplified than in
economics. I have in mind Samuelson’s Foundations of Economics, which
did so much to identify and clarify the small number of principles that lay at
the heart of economic thought as it had evolved in its many branches.
some uncertainty (and a healthy level of skepticism) about just what the
meaning is of abstract systems when used to represent what we understand
of phenomena.
The McCulloch-Pitts model of the neuron and nerve nets provides a nice
illustration of this idea. [Shannon and McCarthy, 1956.] A nerve net is a
system of on-off threshold switches. An external input turns the input
switches on or off or leaves them unchanged. A sequence of messages
(pulses, signals) then flows from unit to unit altering the state of the system
by altering the state in each constituent element. Very simple examples of
such nets exhibit behavior analogous to memory and learning.
Whatever such a system may lack as a model of the brain, it may nonethe¬
less exhibit something profound about the nature of meaning in Langlois’s
systems-referential sense, how meaning emerges from a cascade of intercon¬
nected individuals capable of only elemental stimulus/response behavior and
where every meaning has an exact organizational analog.
CREATIVE INTELLIGENCE
if only they can create. To contrast this property with economic rationality
in the usual sense, I prefer to use the term creative intelligence. If it'did not
exist, we would not.
UNSTABLE SYSTEMS
From the long view, evolution, not “stationarity,” is the fundamental fact of
life. When individuals cannot adapt, they are killed; when organizations
cannot adapt, they are disassembled. When societies cannot adapt, they are
destroyed, swept aside, covered with sand, swallowed by jungle, or
trampled under by a new regime. Humans are destroying virtually all mam¬
mal species with whom they lack a symbiotic tie. Industrialized societies are
destroying the last vestiges of paleolithic man. The internal-combustion en¬
gine has driven away the horse, the oxygen process has eliminated the
Bessemer process. Everywhere and at every level, we see evidence that
economic systems are globally unstable and that in any given form they
eventually stop working. Just now, concern arises about the instability of the
contemporary world economy. The emergent phenomenon that becomes
most crucial for us is the next form of economic policy. From where does it
come?
It must come from the same attribute of mind that guided the subtle hand
of those stone-chippers of so many centuries ago who invented the
spearhead. The attribute that makes it possible to create an image, to set in
place a form in some way unlike any that came before. This attribute creates
a menu of possible ways of life from which policies can be selected and
behavior carried on. Once in existence itself and perpetuated within the
species, it always operates.
When something new is created, since structural information is changed,
its full implications are seldom perceived. The hunting band may have led
some innovative tribe to a flourishing culture, rich in food, ceremony, and
leisure. Eventually, however, when the last woolly mammoth was butch¬
ered, the last herd of giant bison driven over a cliff, a way of life came to an
end. The inventors of fine spears, bows and arrows unleashed a new spiral of
progress, but one that eventually also succumbed in unexpected ways.
The implications are that: (1) Intelligence creates new forms that at cru¬
cial times make possible a continuation of certain living groups; (2) with
changes in structure, a people may persist for a while; (3) the ultimate
consequences of adaption are unknown, and the new system is also (al¬
ways?) unstable, so that new forms must repeatedly be found. Mor¬
phogenetic man is thus an active participant in evolution, matter that has
come not only to reflect on itself, but also to design the new forms it will
occupy in its trajectory through space and time.
INFORMATION ECONOMICS,
TEAM THEORY,
AND AGENCY THEORY
Richard Mattessich
I agree with Langlois when he asserts that “systems theory is not at all
methodological holism in any strong sense.’’ Indeed, in a recent comparison
between Bunge’s important work [Bunge, 1979a] and my own systems
methodology, I expressed this view:
Additional support from the Social Sciences and Humanities Research Council of Canada is
gratefully acknowledged.
625
626 RICHARD MATTESSICH
'The definitions of methodological individualism offered in the two most important standard
works [Dray, 1967, pp. 53—58; Lukes, 1973, pp. 600-601] seem to corroborate this interpreta¬
tion.
INFORMATION ECONOMICS, TEAM THEORY, AND AGENCY THEORY 627
"It seems to be advantageous to make a distinction between information (as evidence) and
knowledge (as hypothesis) and, hence, between information economics and knowledge eco¬
nomics. [Mattessich, 1974a; 1978, pp. 226-233.]
m
alternatives results from selecting the highest among the n expected values
£■(«,), hence, from maximizing the latter over i - 1The modification
or conversion of the decision model into an information-economic model
now consists in making the previously hidden or implicit information signal
yk explicit by substituting the conditional probability p(iS/|;y*)—that is, proba¬
bility of event Sj after receiving message yk—for the previous probability
value pisj). Then the expectations of actions a, (or aok in case of the optimal
action, given signal yk) must also be formulated in terms of signal yk.4
Thus, the expected value of action a,, given signal yk, amounts to the sum
total of the products of the utility u(a„Sj) attributed to each action a, at state
sj(j = 1,,m) multiplied by the conditional probability p(s/|;y*) of state Sj
occurring, provided signal yk(k = 1, . . . , r) is received. The expected value
E(aok,yk) of the optimal action aok, given signal yk, is formed by an analogous
maximization choice. But note that first for each signal there may be a
different optimal action (for this reason, the action must now also bear the
subscript of the signal) and second, that aok need not be identical to a0,
otherwise a systematic information system would hardly be required.
But in IE the emphasis obviously cannot be on the optimal action of a
specific single signal yk but must be on an action optimal under all signal
possibilities. Furthermore, one is interested only in the excess or differential
utility that the information system yields over the expected value of the
optimal action E(a0), that is, compared with a situation employing an ordi¬
nary decision model. Finally, it has to be taken into consideration that the
use of the information system involves some costs C/, either by operating it
(say, in a market-research department) or by buying the information from
outside the firm (say, subscribing to a forecast service). Hence, the expected
net value £> of the information system I can be formulated as the summation
of each optimal expected action for each signal minus the expected value of
the optimal action E(ad), that is, without the employment of any information
system, minus the operating cost Cj of the information system.5 But since
there may be many alternative information systems (I = , N), the
expected net value E0 of the optimal information system must be expressed
as maximum among each of the expected net values Eh that is, for each of
the N information systems feasible.6 This requires that the calculation for Ek
be made for each of the N information systems.7
rn
The economic theory of teams attacks a middle ground. We study the case in
which several persons perform various tasks including those of gathering and
communicating information and of making decisions; but they have common,
not divergent, interests and beliefs. Hence the optimality requirement is easily
defined, just as in the case of a single person. But the single person’s problem
of optimizing his information instrument and its use is replaced by that of
optimizing the allocation of tasks among the members of the team.
8See, for example, Cyert and March [1963] and March and Simon [1958].
9For former attempts and the controversy surrounding them, see Machlup [1967; 1978, chap.
16, pp. 391-423.]
630 RICHARD MATTESSICH
tering into a mutual agency contract that delegates to the agent the manage¬
ment of some kind of entity in compensation for which he or she receives a
portion of the entity’s profit. Depending on the contract, the portion may
range from a fixed amount to a percentage of the profit, with possible penal¬
ties for imperfect fulfillment of the contract establishing some risk-sharing
between principal and agent. Both of them are supposed to work toward a
common goal, but since each of them is trying to maximize his or her own
profit or utility (possibly at the cost of the other party), a conflict of interest
is likely to emerge.
For this reason, an information system is required that informs principal
and agent about the entity’s total profit and supplies further information to
facilitate the agent’s managerial task and the principal’s control of the agent.
As long as the principal can observe the agent’s activities, no moral hazard
arises, and the optimal contract stipulates a fixed remuneration to the agent
with penalties whenever he or she is found to be remiss in his or her duties.
Often the principal’s monitoring of the agent’s activity is too expensive or
cumbersome. In this case, the ultimate profit shares of principal and agent
may not only depend on the total profit, but also on the particular differ¬
entiated information available to each of the two parties.
Applied to the situation of a newly formed company, this means that the
cost of raising equity capital (and the expected return) may be considerably
influenced by the kind of control-and-information system available to share¬
holders. Thus, it may be in the interest of promoters and managers to have
either a voluntary or institutionally enforced system of accounting and audit¬
ing regulations imposed on them, because the lack of directly monitoring
their activity otherwise results in higher shareholders’ risk. Such regulations
could lower the cost of raising capital and increase management’s own por¬
tion of the profit. Hence, the combination of agency theory and information
economics may well be relevant to such practical issues as the ongoing
legislative or quasilegislative activity of the Financial Accounting Standards
Board of the United States.
For these reasons, agency theory fits well into the general program of
information economics, since the value of individual information as well as
the entire information system plays a crucial role in this theory. The theory
also constitutes an obvious link between the information system and the
systemic or organizational structure of the entity involved. Here again, one
encounters rigorous mathematical presentations in the service of an empir¬
ical discipline with the ultimate aim of creating testable or at least refutable
hypotheses.10
10The pertinent literature is vast, and the bibliography arranged by Demski and Feltham [1981]
comes close to a thousand articles and books. But to be more specific, I list the following
relevant publications (beyond those previously indicated): Atkinson and Feltham [1980]; Arrow
[1974]; Barnea, Haugen, and Senbet [1981]; Dickson, Senn, and Chervany [1977]; Fama [1980];
Hilton [1981]; Holstrom [1979]; Itami [1977]; Mattessich [1975; 1983 forthcoming]; Ng and
Stoeckenius [1979], and Zmud [1979],
ON THE RECEPTION OF NOISE
A Rejoinder
Richard N. Langlois
631
632 RICHARD N. LANCLOIS
farmer” was all about, and his discussion is a successful if slightly jar¬
gonized expansion of what I was trying to say. Also, his treatment of the
important and often-neglected distinction between formal and informal or¬
ganizations is excellent, even if, with Friedrich Hayek, I would prefer to
avoid confusion and call the latter not an organization at all but an “order.”
[Hayek, 1973, pp. 36-40.]
Beam asks whether I would concur that models of social interaction
should involve both “the science of communication and the science of trans¬
action.” Absolutely. The economics of information, in particular, should be
precisely a study of both communication and transaction. I am a little trou¬
bled, though, with the idea that the analysis of transactions should rest
heavily on the concept of power. If anything, it seems to me, it should be the
other way around: Power, insofar as it appears as a social phenomenon, is
something that needs to be explained in terms of the structures of transac¬
tions involved. Economics has lately turned renewed attention to the idea of
the transaction as a unit of analysis, a trend that has opened up some inter¬
esting new areas of inquiry. [O. E. Williamson, 1979; 1980.]
Now, let me turn to those who found more fault with my presentation.
Both Buckley and Richard Mattessich are still bothered by an issue I was
trying to defuse: whether systems theory differs from methodological indi¬
vidualism in its desire to consider the whole greater than the sum of the
parts. I probably should have realized that my negative answer would raise
some hackles; there is much intellectual capital invested in the proposition
that systems theory is less reductionistic than methodological individualism
and more attentive to emergent phenomena. There may, in fact, be subtle
distinctions to be worked out in principle, but the important issues will
continue to lie elsewhere.
Systemism and methodological individualism are not (necessarily) identi¬
cal. (I had thought I was clear on this point.) But what distinguishes them is
not a debate about whether properties of the whole can be deduced solely
from properties of the parts; rather, it is a debate about what shall be consid¬
ered the parts. This way of putting things should raise even more hackles.
For the methodological individualist’s response—that in the social sciences
the only acceptable parts are the knowledge, expectations, intentions, and
actions of human beings—calls into serious question whole classes of fash-
action is to illustrate the extent to which it cannot be so reduced in more complicated systems.
Some of the confusion may have arisen from my portrayal—for heuristic reasons—of the
decision-making farmer as a real person. What I was actually trying to illustrate was the ideal
type of an economic agent as portrayed in a mathematical maximization model. In such a
model, meaning is indeed “the mapping between the learned properties of rain and the range of
possible consequences of rain for the farmer’s particular goals and interests at the moment”;
but in the model farmer, this is exactly a mapping of acreage into expected profit, because the
way in which I have constructed the model rules out any changes in internal states or values and
goals. I heartily agree that this is too narrow a definition of meaning in economics: That is one of
the central points of my paper.
RECEPTION OF NOISE 633
3 Buckley also seems to confuse explanation and prediction in this way when he writes that “we
cannot begin to explain or predict the properties or behaviors of complex dynamic wholes from
a knowledge of the parts. ...”
4For a good discussion of exactly this hydrogen and oxygen example, see Dennis C. Phillips
[1976, pp. 14-15, 32-33],
5It is in this explanatory sense that the social sciences, as Buckley suggests, should be synthetic
rather than analytic. But Anatol Rapoport was not the first to argue this position. In words first
published in the 1940s, Hayek wrote that: “While the method of the natural sciences is . . .
analytic, the method of the social sciences is better described as compositive or synthetic.”
(Hayek, 1941 and 1979, pp. 66-67.)
61 have referred to Robert Costanza’s article as an example because it was certainly considered
system-theoretic in some circles, as evidenced by the fact that it was reprinted in the General
Systems Yearbook series.
634 RICHARD N. LANGLOIS
The impact of man in the evolutionary process arises because of the capacity of
his images—that is, the knowledge present in his mind—to grow by a kind of
internal breeder reaction: the imagination. It is this which has given the human
nervous system such a fantastic social-evolutionary potential, a potential of
which we have probably hardly used up one per cent in the brief history of the
human race. (Boulding, 1964, p. 141.)
RECEPTION OF NOISE 635
In fact, what I was trying to say is something rather different. What Smith
was getting at in his conception of an “invisible hand” is not the idea that
given an appropriate set of legal institutions, the market will produce homeo¬
stasis governed by negative feedback. What Smith argued was that given an
appropriate set of institutions, the free exchanges of individuals will lead to
the production of wealth—to continual and cumulative change in organiza¬
tion and technology. [Bohm, 1982.] For Smith, the unintended conse¬
quences of individual action within what he called “the system of natural
liberty” would be in the nature of evolutionary morphogenesis. [Adam
Smith, 1776 and 1936, p. 651.]
As I tried to emphasize in my paper, it is important to get straight which
kind of a system one thinks the economy is. If one does not conceive of it as
an evolving order but as a cybernetic system in need of central direction,
then one is likely to interpret change and innovation not as morphogenesis
but as chaos and disorder.
But the choice of which lens one uses to view the economy (or society
more generally) is not arbitrary. As a historical matter, Buckley is dead
wrong when he says that Western economic systems have been generally
feeble and unstable. It may or may not be true that the economicopolitical
system is unstable. The free institutions on which the market economy is
based are the exception not the rule in history. But when those institutions
were more or less firmly in place—as in Western Europe and America during
most of the last 200 years—the result, as even Marx recognized, was the
greatest outpouring of human creative intelligence and energy in the history
of mankind. There were certainly recessions and depressions (some of
which were arguably caused or aggravated by attempts at central control),
but it is a perversity of perspective to call the system feeble or unstable.7
71 was pleased in this context to see Churchman point out the fallacy of seeing world poverty
and starvation as basically a food problem. The problem is an institutional one—a problem of
setting up in Third World countries the kinds of social and economic institutions conducive to
the production of wealth. But I was distressed at his explanation in terms of superp°wer
politics, although the difficulty may have arisen from the brevity with which he expressed
himself. As stated, his explanation is an instance of what Karl Popper called “the conspiracy
636 RICHARD N. LANGLOIS
theory of society,” a mode of social explanation more primitive than most forms of theism.
[Popper, 1966, pp. 94-95.] The reason this mode of explanation fails, as Popper notes, is that we
live in a complex world in which the prevalence of unintended consequences makes power a
complicated affair and a generally implausible explanation. It is not clear that countries would
all know what to do—or could do it if they knew—in the absence of superpowers. Neither is it
clear that superpowers can prevent nations so determined from setting up wealth-producing (or
at least starvation-preventing) systems. There are a number of countries (Taiwan and South
Korea) that are certainly the focus of superpower attention but have nonetheless been able to
set up fairly successful wealth-producing institutions. And there are dozens of forgotten coun¬
tries entirely unable to change their behavior patterns in the face of starvation.
8The political order Smith advocated was never literally laissez faire; it was, in fact, based on a
common-law framework involving property rights and a strong conception of justice. His claim
was that when these institutions are in place, people left to themselves will produce good social
outcomes. In more system-theoretic terms. Smith was arguing that good social outcomes would
be the emergent result of individual action within the constraints of certain boundary condi¬
tions. If the catch-phrase laissez faire meant anything, it meant that the economic system should
not be viewed as requiring external central control in the manner of an aircraft or spaceship.
This is the control-theoretic content of the old slogan “a government of laws and not of men.”
Moreover, Smith’s advocacy of the “system of natural liberty” was based on its justice proper¬
ties; that it led to good economic outcomes was for Smith simply a happy coincidence.
[Buchanan, 1979.] The justice properties of Smith’s system, incidentally, are closely related to
those implied in the Kantian imperative cited by Churchman. I was very pleased to see this
reference to Kant, but I was also a little surprised to see this imperative discussed favorably in
an article extolling rationalist planning—since, to most philosophers, the Kantian imperative is
taken as implying individual rights that, if not exactly inconsistent with rationalist planning, are
nonetheless conceived of precisely as strong constraints to, or protections against, the central¬
ized implementation of rationalist plans. [Nozick, 1974; Fried, 1978; Dworkin, 1978.]
RECEPTION OF NOISE 637
Fritz Machlup
The pronoun “They” in this heading does not refer just to the members of
the diverse cultures interested in the study of information but it refers to all
who use the word. If a scholar adopts a word from everyday life to denote
This manuscript was unfinished at the time of Fritz Machlup’s death; it was prepared for
publication by Una Mansfield, to whom he had handed over the completed parts. Missing
subsections are indicated by an Editor’s Note at the point where they would have appeared in
the section “What They Mean By Science.” Fortunately, the subjects missing from this text
have been treated elsewhere by Machlup, and scholars interested in knowing more of his views
should consult the volume Knowledge and Knowledge Production. [Machlup, 1980.]
641
642 FRITZ MACHLUP
some specific “designata” different from the common meanings, the latter
are not thereby extinguished. They remain in the dictionary and we have no
right to disregard them.
Dictionary Definitions
1 Lest the slightly archaic word “to impart” is misunderstood, its definition is “to give a part or
share of; to bestow, communicate as knowledge or information; to make known, tell, relate
. . . .” Hence, it always refers to an activity, not just an opportunity of being observed. Learn¬
ing about something by looking should not be confused with learning by being told.
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 643
2This third distinction has been most interestingly elaborated by Boulding: “. . . we cannot
regard knowledge as simply the accumulation of information in a stockpile, even though all
messages that are received by the brain may leave some sort of deposit there. Knowledge must
itself be regarded as a structure, a very complex and frequently quite loose pattern, . . . with its
parts connected in various ways by ties of varying degrees of strength. Messages are continually
shot into this structure; some of them pass right through its interstices (. . .) without effecting
any perceptible change in it. Sometimes messages ‘stick’ to the structure and become part of it.
. . . One of the most interesting questions in educational theory and practice is under what
conditions does information ‘stick’ in this way, and under what conditions does it fail to ‘take
hold’ of the structure of knowledge. . . . Occasionally, however, a message which is inconsis¬
tent with the basic pattern of the mental structure, but which is of such a nature that it cannot be
disbelieved hits the structure, which is then forced to undergo a complete reorganization.”
(Boulding, 1955, pp. 103-104.)
Manfred Kochen made a statement about “[w]ays to integrate the growing body of informa¬
tion into a commonly shared fabric of knowledge.” (Kochen, 1970, p. 44.) This may sound as if
it were the same as Boulding’s view but really is not. Kochen evidently thinks of a social “body
[stock] of information” and a social “fabric of knowledge,” as if the sum of individual minds
could be hypostatized into an “aggregate mind” of the whole society, or perhaps an “electronic
encyclopedia” that includes all fragmented and obsolete piecemeal messages that once con¬
veyed timely information of transitory relevance. There is no growing body of information, for
what is not integrated into knowledge is thrown out or forgotten. There is, of course, a growing
flow of information and some of it is put into storage in computer systems. Whether the eventual
output is information or knowledge depends on the individual users.
644 FRITZ MACHLUP
3In an earlier discussion of the “Uses, Value, and Benefits of Knowledge,” I used the analogy
of a transport and delivery service in relation to the object transported and delivered: “To use
information—a process, mind you—is to listen, to look at, to read; in short, it is its reception
and, if possible, the full or partial understanding by the recipient. The use of the knowledge
conveyed is something else. The act of delivering is one thing, the object delivered is another
.... Use of mode of transportation and use of the transported object are different things.
Likewise, use of a mode of information should not be confused with the use of the message or
knowledge conveyed.” (Machlup, 19796, pp. 63 and 65.)
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 645
services. On the other hand, we would frown on education programs that fill
the student’s head with loads of information: We want them to disseminate
knowledge of enduring value and to develop a taste or thirst for more knowl¬
edge, not just information.4
4The last paragraph reproduces what I wrote in chapter 1 of volume 3 of my projected ten-
volume series. [Machlup, 1984.]
646 FRITZ MACHLUP
rethinks, interprets, infers. Only when he tells his fellow scientists about his
work—by word of mouth, by a written report, or by a published paper—is
genuine information (in the original sense of the word) being conveyed.
To use the words information and observation as synonyms is bad lin¬
guistic practice; it leads to confusion between different sources of knowl¬
edge and is therefore epistemologically and methodologically unsound. The
fact that eminent scientists have sometimes been semantically imperceptive
and have metaphorically used the term information in lieu of observation is
no good reason for emulating them.5
The use and misuse of the term data is due, in part, to linguistic ignorance.
Many users do not know that this is a Latin word: dare means “to give”;
datum, “the given” (singular); and data, “the givens” (plural). Data are the
things given to the analyst, investigator, or problem-solver; they may be
numbers, words, sentences, records, assumptions—just anything given, no
matter in what form and of what origin. This used to be well known to
scholars in most fields: Some wanted the word data to refer to facts, espe¬
cially to instrument-readings; others to assumptions. Scholars with a hy-
pothetico-deductive bent wanted data to mean the given set of assumptions;
those with an empirical bent wanted data to mean the records, or protocol
statements, representing the findings of observation, qualitative or quantita¬
tive. With this background of historical semantics, a reader of recent
definitions of, or statements about, data cannot help being appalled.
There are writers who insist that data consist entirely of numbers.6 That
this restrictiveness is (fortunately) exceptional can be inferred from the fact
that the majority qualifies the noun and speaks of numerical data and statisti¬
cal data where the givens are sets of numbers. Such numbers are rarely
given in nature, but instead have to be produced by employing sophisticated
techniques using instruments that sometimes are built on the basis of rather
complex theories. A good deal of prior knowledge has gone into the con¬
struction of measuring devices that are to furnish the numerical data wanted.
Yet, we often read of “raw data” being used to produce information and
5Among the metaphoric users is Claude Bernard: “We must observe without any preconceived
idea; the observer’s mind must be passive, must hold its peace; it listens to nature and writes at
its dictation." (Bernard, 1865 and 1927, p. 22. Emphasis added.) The statement is wrong in
more than one respect, but especially in the phrase I emphasized.
As a curiosum I may cite a phrase that compounds the confusion by speaking of “macro-
scopically observable information,” where “information” evidently stands for “phenomena.”
The same writer “informs” us wrongly that “entropy is related to ‘missing information’ ”; what
is missing is the possibility of observation. [Shannon and Weaver, 1949.]
6“The corpus of information . . . consists of two types of information—non-data and data.
Non-data is non-numeric .... Data, on the other hand, is numeric, highly formatted and results
from analysis.” (Dolan, 1969, p. 41.)
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 647
eventually knowledge.7 Perhaps raw data are only relatively raw, in that
they are inputs for the production of more highly fabricated kinds of infor¬
mation.8 Many writers prefer to see data themselves as a type of informa¬
tion, while others want information to be a type of data.9
One can probably find quotations supporting all possible combinations of
the three terms or of the concepts they are supposed to denote.10 Each is said
to be a specific type of each of the others, or an input for producing each of
the others, or an output of processing each of the others. This semantic
muddle, however, need not cause any serious trouble, because the argu¬
ments in which data, whatever they are, play a central role are relatively
simple: Data entry, data storage, data retrieval, data processing, data ser¬
vices, and all the rest, refer simply to things fed into a computer. These
things, now data from the point of view of the programmers, operators, and
users of the computer, need not be data in any other sense. Drafts of a
manuscript for a learned monograph (or of a mystery story, for that matter)
may have been typed into the computer; or the subject index for a textbook;
a bibliography of writings on the history of French painting; detailed statis¬
tics of the gross national income of the United States from 1940 to 1980;
expert knowledge for the diagnosis of diseases of various kinds; graphs and
images of all sorts; or what not. Most of these inputs, now in the memory of
the computer system, are very far from being raw data in the sense of
empirical scientific analysis. Whether they are pieces of fragmented informa¬
tion—say, the football scores of American college teams for Saturday,
November 13—or whether they are elaborate compendia of systematic
knowledge in some discipline, for the users of these materials accessible in
the computer memory they are data stored and data retrieved.
Of course, these materials are not stored in the form in which they were
typed, with words and numbers; instead, they are encoded by means of
digital data representation. Data transmitted from a computer are “repre¬
sented by the presence or absence of an electrical impulse (representing a bit
‘on’ or a bit ‘off’)-” (Shelly and Cashman, 1980, p. 86.) Thus, there may be a
slight source of semantic confusion between data and data representation,
7Thus we read about “transformations that result in data becoming information and knowl¬
edge.” (Kent, 1982, p. 315.)
8For example, information is defined as “the result of processing of data, usually formalized
processing.” (Hayes, 1969, p. 218.)
9Manfred Kochen cites approvingly a statement he attributes to West Churchman: “Informa¬
tion ... is essentially raw data. Knowledge is interpreted data.” (Kochen, 1970, p. 48.) Thus
both information and knowledge are said to be data; and information is said to be raw, though in
fact it may be thoroughly processed and well-done.
10 “In more elaborate versions of the . . . model the analyst can request the synthesist to
synthesize a new CKS [contingent knowledge structure] . . . that may change the effective
information that the analyst obtains from the raw input data. . . . The contingent knowledge
structure (CKS) is a data structure which represents the understander’s knowledge of the state
of the world.” (R. J. Bobrow and Brown, 1975, pp. 115-117.)
648 FRITZ MACHLUP
but this can easily be avoided by using terms such as impulse or signal, or
even symbols.11
Why all this has caused so many writers in information science to stumble
over their attempts to define data and information in terms of one another is
probably explained by their ambition to sound more sophisticated than the
problems warrant. There is no need to establish either a hierarchy or a
temporal sequence in discussing data and information. Apart from computer
systems, the two words may be equivalents—say, “we have the informa¬
tion” or “we have the data needed for a particular inquiry”—or they may be
otherwise related—say, “our data do not furnish the information wanted in
this case.” On the other hand, in a computer system, data are what has been
fed into the memory of the system and is now available for processing.
This oversimplified presentation on “what they mean by data” has disre¬
garded some differences of possible importance. Take, for example, three
possible outputs: (1) a printout that gives us exactly what has been fed into
the memory of the computer, the same words and numbers in just the order
in which they have been stored; (2) a new arrangement of the data, after a
programmed process of sorting and re-ordering of the stored materials
(chronological or alphabetical ordering, or selecting on the basis of detailed
instructions); and (3) an output different from the stored data as a result of an
analysis made by the computer in accordance with involved instructions
specified, step by step, in a highly sophisticated program.12
Since the writers on computer and information sciences are arguing about
the question whether the results of such high-degree processing should still
be called data or should be referred to as information, 1 submit that—in this
case—the name does not make a bit of difference for any reasonable pur¬
pose. From the point of view of the computer operator, one may still say the
data are in the memory and the output is something else. From the point of
view of the user of the output, data and information are equivalent terms.
The people selling management information systems (MIS) feel better if they
call the output of their system information, that is, something of a higher
order. Some people in management-decision theory prefer to say that the
MIS output is not yet all that is needed for a decision, that more analysis
and, especially, more judgment is needed, and hence that the output sup¬
plied to the decisionmakers is still in the “lower” category of data. I repeat
that this quibbling is of no consequence.
As a relief from the dreariness of this discussion, I want to end it with a
quotation from the paper by Jesse Shera in this volume. He exclaimed:
“Data, data everywhere and not a thought to think.”
With this business of the semantic or operational links between data and
information out of the way, we may return to our survey of meanings of the
word information. We shall find that numerous proposals for restrictive
definitions have been made and, for better or worse (more often, for worse),
widely adopted.
There are those who link, by definition, information to decision-making
and action.13 This is one of the many instances in which a word that has a
wide meaning is appropriated for use with a very narrow meaning. For these
definers, information, to be information, has to have value, has to be used
for decision-making, and has to be designed to lead to action. I submit that
more than 90 per cent of all information received during a day (week, month,
year) by people in all walks of life is not related to any decisions or impend¬
ing actions. Just think of the sports sections of newspapers and broadcasts,
or of the general news reports day after day! Most recipients of such infor¬
mation do not entertain any thought that it may help them in making deci¬
sions for practical actions. Even on the highest level of thinking, information
need not be in the service of pragmatics.
Another restriction built into proposed definitions of information is that it
must reduce uncertainty on the part of those getting informed. This issue is
complicated by the fact that uncertainty means different things in the discus¬
sion of physical phenomena, processes, and measurements on the one hand,
and in the discussion of human behavior and cognitive processes, on the
other. Since I am still talking about information in the sense of telling some¬
thing to somebody, I may at this point confine myself to uncertainty as it
bears on persons receiving information and to the definitional contention
that information is not information unless it reduces uncertainty. This re¬
striction by definition has been imposed by eminent representatives of the
behavioral and social sciences, especially psychologists and economists.
[Garner, 1962; Arrow, 1979.] For the moment I need not say more than that
this definition is at variance with the common usage of the word. Countless
13 “Information is data of value in decision-making.” (Yovits and Abilock, 1974, p. 163.) The
perpetrators of this definition proceed to the statement that: “The information contained in a
decision state is related to the mean square variance of the expected values of the courses of
action.” (Ibid., p. 166.)
650 FRITZ MACHLUP
Students of linguistics use the term information not always for what one
person tells another by some form of explicit message but rather for ex¬
plaining the relations among words in a sentence. In this formulation, infor¬
mation, in a very special sense, is something implicit in verbal expression; it
is not an action, process, or piece of knowledge, but it is what identifies the
meaning of words. This usage has the interesting result that meaning is
defined as information whereas, ordinarily, information is viewed as convey¬
ing meaning and as having a “meaning-content.” This apparently contradic¬
tory interrelation is duplicated in another instance: A message is often said
to contain information, and information is said to consist of messages. One
way to overcome such inconsistencies may be to distinguish information
from representations of information and from transmission of information.
In order to avoid confusion between the transmission of physical signals
and the meaning-content of messages or statements, Bar-Hillel and Carnap
adopted the term semantic information. As if there could be such a thing as
nonsemantic information! Yet, it was probably necessary to use that glaring
pleonasm in conferences with representatives of the statistical theory of
signal communication who had appropriated the word information to denote
“signal apart from meaning.” Bar-Hillel and Carnap, emulating the com¬
munication theorists’ efforts at measurement, proceeded to search for a unit
of semantic information; after analyzing the notion of an atomic statement
they proposed a unit of measurement, called semantic content-element.
[Bar-Hillel and Carnap, 19536, pp. 148-149.] In any case, for these philoso¬
phers of language and inductive (probabilistic) inference, information is
defined in terms of semantic content-elements of verbal statements.14
For many logicians and philosophers, information consists of statements,
and the amount of information contained in a statement is determined by the
relative number of excluded alternatives. If a statement predicates one thing
and thereby denies or excludes many alternatives, it is said to constitute
more information than a statement that excludes only a few other possible
states of the world. These considerations may make good sense in scientific
testing, where alternative outcomes are not too numerous; but in primitive
pieces of news the number of alternatives excluded by the statement may be
indefinite. It would depend on the expectations the recipient of the informa-
All meanings and definitions just reviewed were of information in the general
sense of something being told to somebody, where this somebody was sup¬
posed to grasp what was being told. In other words, it was meaningful
(semantic, cognitive) information; either it was the act (or process) of telling
or it was that which was being told. Some definitions, however, were restric¬
tive, refusing the designation information to some types of (what others call)
information. Thus, to list the restrictions most often proposed, information
is not recognized as information according to (any particular) proposed
definition unless (1) it is about something previously unknown to the recipi¬
ent; or (2) it is about something previously less assuredly known to the
recipient; or (3) it affects the stock or structure of the recipient’s knowledge;
or (4) it consists only of raw data, not yet interpreted; or (5) it is useful to the
recipient in some way; or (6) it is used in decision-making by the recipient; or
(7) it bears on actions contemplated, considered, or actually taken by the
recipient; or (8) it reduces the recipient’s uncertainty; or (9) it helps identify
the contextual meanings of words in sentences; or (10) it excludes some
alternatives to what is predicated in a statement; or (11) it changes some
belief held by the recipient, particularly with respect to the distribution of
probabilities in the recipient’s view.
This is surely not an exhaustive list of alternative definitions of informa¬
tion directed to a human mind; but it suffices to show that many analysts
prefer to make things true by definition and to save words to modify a more
broadly defined term. For example, instead of saying new information or
unexpected information, some choose to omit the adjective and decree that
novelty or unexpectedness is a necessary characteristic of information.
We now turn to metaphoric uses of the word. We shall, in the next
subsections, discuss uses of the term information in connection with de¬
scriptions or models of processes or phenomena pertinent to living humans
or, more generally, “living systems.”
652 FRITZ MACHLUP
Not all writers on nervous systems deem it necessary to use the term infor¬
mation in describing what goes on when sensory impulses or signals are
conveyed to the central nervous system or when motor controls are exer¬
cised by it over muscles or other organs. In scanning some of the texts and
handbooks of neurophysiology I have found that some writers can do quite
well without the term information; perhaps they have decided that it is better
to avoid the (often misleading) metaphor.15 Other writers, however, seem to
have concluded that the nervous system can be more easily explained if
communication systems such as telegraph and telephone are taken as
analogies.16 From analogy to metaphor is only a short step and users of a
metaphor may become so used to it that they adopt the term as the genuine
designation of the phenomena or processes they wish to describe or explain.
It is not always clear whether the writers on sensory physiology, psychophys¬
ics, and cognate subjects want the word information to refer to its content
or only to the process of transmission. When neurophysiologists describe
nerve-communication in terms of firings from one neuron to the next, they
can easily avoid references to “information being passed on.” However,
they have a much harder time when they want to describe tactile sensations
from mechanical stimulation being picked up by receptors in the skin and
sent on to reach a central neuron. Is the pin-prick or the tickle a stimulus, an
impulse, or perhaps a piece of information encoded by cutaneous receptors
and decoded by central neurons? To judge from the literature, it seems to be
difficult to dispense with the use of the term information.17
15 “Neurophysiology deals with functional aspects of the nervous system, with transmission of
nerve impulses, motor control, reflexes, and even perception, emotion, and mentation.” (Ochs,
1965, p. 9.) “[T]he means of propagation of the nerve signal itself is electrical in nature.” (Ibid.,
p. 11.)
16“The nerve fibre can be compared, from the point of view of its function, with a cable carrying
information.” (Zimmermann, 1978, p. 68.) This writer proceeds to discuss information trans¬
mission in the nervous system and to employ the term information not only for the transmission
process but also for measuring the amount of information in the sense used in the mathematical
theory of communication.
17 Analysts of the sensibility of the skin distinguish among several types of “receptors of cutane¬
ous sensations” and report on mechano-reception (the skin’s sensibility to touch and pressure),
thermo-reception (the skin’s sensibility to heat and cold), and nociception (the skin's sensibility
to pain). [R. F. Schmidt, 1978, p. 81.] The process of transmission from the receptor to the
receiving end of the neural string—yes, the receptors in these instances are the senders—is
similar for different sensations. The question of just what is being transmitted is difficult to
answer; one can understand the analysts’ terminological promiscuity in the exposition of vari¬
ous sensibilities, when they rather indiscriminately dally with impulses, stimuli, sensations,
inputs, signals, and information. In his discussion of vibration receptors in the skin, Robert
Schmidt states that “[t]his receptor can transmit no information about the depth of skin indenta¬
tion.” (Schmidt, 1978, p. 91. Emphasis added.) But “mechano-receptors . . . can signal the
position of a joint....” (Ibid., p. 96. Emphasis added.) However, “signals” deriving “from the
central motor systems . . . appear to send a ‘memo’ (efference copy . . .) of their signals to the
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 653
Perhaps we should not worry so much about proper word use, and accept
the fact that information has become an all-purpose weasel-word. Even a
writer as careful and lucid as Michael Arbib, in a passage in this volume,
manages to use the word information in three senses: as a process, as a
current (specific) content, and as the accumulated content of several previ¬
ous messages.18 Lest the reader get the impression that the metaphoric use
of the term information in neurophysiology began only with the emergence
and dramatic growth of the theory of information, I should mention that
many earlier instances can be found in the literature.19
Brain research is another subdiscipline in which some writers are fond of
using the term information while others seem to avoid it. We find this differ¬
ence in linguistic preference in different reports on the split brain and the
“interhemispheric communication” between the two halves of the brain.
[Sperry, 1965, 1982; Gazzaniga, 1970, 1981.] Thus, we read that after sever¬
ing the connections between the hemispheres “information learned by one
half-brain did not transfer to the other.” (Gazzaniga, 1981, p. 517.) The two
halves were no longer on speaking terms, so to speak.
Before I proceed to the next group of habitual users of the term informa¬
tion in biology, the geneticists, I may point to an interesting difference in
metaphoric use. Writers on neurophysiology are essentially concerned with
what once was discussed in terms of stimulus and response and is now
discussed in terms of signals and information; the analogy here refers to brief
ad hoc messages, telling in a few spoken words, as it were. On the other
hand, writers on genetics, with their models of a genetic code, refer to long
sets of long-term instructions, specifications with a blueprint for future per¬
formance, hence, telling in a long written scroll. Thus the metaphor “infor¬
mation” alludes to “spoken messages” within the nervous system, but to
muscles, to interact centrally with the sensory input from proprioceptors. These efference
copies give advance information about the intended muscle activity and the movements that
will result.” (Ibid., p. 99. Emphases added.) Even more metaphorically, “Pain . . . informs us
of threats to our bodies, for it is activated by noxious (tissue-damaging) stimuli.” (Ibid., p. 111.
Emphasis added.)
18In his paper in Section 2 of this volume, Arbib has two sentences where all three senses are
alluded to: “Information picked up modifies the perceiver’s anticipations of certain kinds of
information that, thus modified, direct further exploration and prepare the perceiver for more
information.” Arbib admits, however, that we do not know yet “how the organism can be
committed to an overall action by a population of neurons, none of which has global information
about which action is appropriate.” Global information, we assume, refers to accumulated
experiences (perhaps knowledge).
19 “We know that when the eye sees, all the consequent information is transmitted to the brain
by means of electrical vibrations in the channel of the optic nerve .... The impulses which
flow in the arm nerves of a typist convey to her fingers the translated information which reaches
her eye or ear, in order that the fingers may be caused to strike the proper key.” (Bush, 1945, p.
58.) I understand that more recent descriptions would not be in terms of electrical vibrations,
but the point in offering the quotations was to show the early use of the word information for
neurophysiological signals.
654 FRITZ MACHLUP
20For example, after talking about the documentary sense of information, Bertram C. Brookes,
a distinguished writer in information science, made this pronouncement: “Clearly, they [the
geneticists] are not using ‘information’ in its documentary sense and it is a far cry from docu¬
mentary ‘marking’ and ‘parking’ to the biochemical processes of a living cell. But while the
information analogy is helpful to microbiologists I see no reason why it should not be used.”
(Brookes, 1974, p. 146.)
21 The genetic code is made up by four nucleotide bases (symbols, characters, letters), conven¬
tionally denoted by the letters A, T, G, and C. Pairs (or doublets) of letters would allow coding
of only 42, or 16, “words,” which would be insufficient to specify 20 amino acids. Sequences of
three letters (triplets) allow coding of 43, or 64 words, or codons, and thus permit a good deal of
redundancy. The “sequence of nucleotides in DNA carries the genetic information that orders
(codes) the sequence of amino acids in proteins.” (Watson, 1965, p. 99.) “The molecule’s
message written in nucleic acid code” may represent structural genes, controlling “how
molecules of specific amino acids are combined to form a particular protein, programming the
pattern of spatial organization of that protein.” Other messages written in the same code
represent regulatory genes, programming “the pattern of temporal organization, determining
when syntheses of protein molecules shall occur by controlling rates of synthetic processes.”
(J. G. Miller, 1978, p. 221. Emphases in the original.)
The allusion, in the text above, to punctuation marks refers to termination or the end of a
chain of triplets. “Termination is not a random process, but is highly controlled by special ‘full-
stop’ or termination codons. Recent studies have shown that certain triplets or codons (called
‘nonsense codons’) in the mRNA [messenger RNA] chain automatically bring about termina¬
tion of the peptide chain at that point. The nonsense codons are UAA, UAG, and UGA. When
the ribosome reaches a nonsense codon, the bond between the final amino acid and the tRNA
[transfer RNA] molecule to which it is attached is hydrolyzed.” (Baker and Allen, 1982, p. 561.)
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 655
22For James Miller, “A society is a large, living, concrete system with organizations and lower
levels of living systems as subsystems and components.” Miller explicitly rejects Toynbee’s
view of looking at society as “a product of the relations between individuals,” and likewise the
definition by Parsons, Shils, and others, of a social system as “interactive relationships of a
plurality of individual actors.” (J. G. Miller, 1978, p. 747.) Here is one of Miller’s explicit
statements on information (or perhaps a proposed definition): “Information is the patterning of
matter-energy in systems.” (Ibid., p. 1030.)
656 FRITZ MACHLUP
did not confuse it with political individualism—and who rejected the myths
of romantic holism—not to be confused with the holistic aspect of general
system theory. Some of the papers in Sections 8 and 9 of this volume deal
with the problems in question and generate heat and light in the process.
(The reader may enjoy both these outputs.)
What the word information means to individual members of groups, or¬
ganizations, and society, and also to analysts of these social systems, is not
different from what the dictionary definitions say. On all these levels, infor¬
mation has or is a meaningful content. In the analysis of information flows,
of the possible relations between information and decision, of possible ef¬
fects of information upon mass action, and of several other problems, spe¬
cialists in the study of organizations may emphasize other aspects than do
the students of economic society or political society; some writers may find
it expedient to make different specifications with regard to degrees of uncer¬
tainty or changes in beliefs, or with regard to revisions of expectations or
even of goals; still, the fundamental notion of information is the same in all
social sciences.
There have been writers who were fascinated by notions of social infor¬
mation and social knowledge, in the senses of new information acquired or
total knowledge possessed by society—where society was supposed to be
something quite different from the individuals that compose it.23 The basic
ideas of system-thinking seem to foster a belief in the reality of a whole apart
from its components and their interactive relations. If information is some¬
thing that reaches a mind, or several or many minds, it ought to be clear that
the whole does not have a mind of its own and can neither receive nor
process the information that has reached its members. To be sure, some
special decisionmakers, say, members of legislatures, administrations,
public agencies, and so forth, may have the right and the power to act upon
received information (collectively) issuing laws, decrees, and regulatory or¬
ders controlling, constraining, inducing, or prohibiting certain kinds of ac-
23Norbert Wiener has in his Cybernetics a chapter on “Information, Language, and Society,”
in which he tries to differentiate the “intelligence of society” or “community” from the intelli¬
gence of its members. [Wiener, 1948 and 1961, p. 162.] He states that there is “no necessary
relation in either direction between the amount of racial, or tribal, or community information
and the information available to the individual.” (Ibid., p. 158.) Not all references, however, to
social and individual knowledge are naive or fallacious. On sensible distinctions between social
and individual knowledge, proposed by Bertrand Russell, Alfred Schutz, Thomas Luckmann,
and others, see my book on Knowledge and Knowledge Production. [Machlup, 1980, pp. 161 —
162, 167-173.] However, the contributions of system theorists to the flows of information into
the system and among its components are, by and large, rather confusing. Thus, for example,
James Miller speaks of a “special form [of information flow], the flow of money, which is one
sort of information.” (J. G. Miller, 1978, p. 1027.) Miller may possibly be thinking
of prices emerging in commodity markets and interpreted as giving signals or information to
producers and consumers. But this has nothing to do with the circulation of money; and it surely
does not support the notion that flows of money are flows of information. It is not a useful
analogy—just a fallacy.
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 657
tions. Still, both the controllers and the controlled are individuals, even if
the controllers may claim to be acting on behalf of society. Moreover, mass
actions by individuals not entitled or empowered to legislate or to govern
may well exercise powerful influences on governors and controllers. The
titles and constitutional functions assigned to particular components of a
social system do not necessarily determine their effectiveness; the control¬
ling forces may be widely diffused, independent of any official organization
chart.
24One of the few specialists who was sufficiently perceptive to see the difference was Colin
Cherry: “All communication proceeds by means of signs, with which one organism affects the
behavior [or state] of another. ... In certain cases it is meaningful also to speak of communica¬
tion between one machine and another as, for example, the control signals which pass between
a guided missile and a ground radar.” Cherry proceeds to consider the question how to “distin¬
guish between communication proper, by the use of spoken language or similar empirical signs,
and other forms of causation.” (Cherry, 1957, p. 219.)
658 FRITZ MACHLUP
voluntary versus semiautomatic. “If I push a man into the lake, he inevit¬
ably goes in; if I tell him to jump in, he may do one of a thousand things.”
(Cherry, 1957, p. 220.) To be sure, the question whether X was a push or a
sign may be of little interest to a scientist who wants to explain the state Y of
the man, dead or alive, but surely wet. However, to a scientist not satisfied
with algebra it will make a difference whether X was a physical force or
rather a sequence of signs in gestures or words. Indeed, some scientists may
wish to inquire what kind of meaningful information has induced our man to
jump in the lake: a command, an advice, a threat of worse consequences, a
bet, an offer of a bribe, a report of a sad event? If any of these messages and
any acts of physical force can be indiscriminately called information, this
term has lost much of its usefulness.
Despite this grave charge, I concede that, with the degree of discernment
that one should expect from any academically trained person, the probabil¬
ity of confusion caused by metaphoric uses of the term is not intolerably
great.
We now come to the last group of users of the term information; it is a rather
special case: They use information and amount of information in a sense that
has so little to do with any traditional or metaphoric meanings of the word
that one can only wonder why the scientific community has allowed it to
continue. In the Prologue to this volume, especially in the subsection on
“Information Theory,” it was pointed out why the misnaming has been
infelicitous, misleading, and disserviceable, and also why frequent attempts
to correct this state of terminological affairs have been unsuccessful. At the
present juncture I merely wish to state just what the information theorists (or
rather the expositors of the mathematical theory of communication or signal
transmission) meant when they said information.25
Information in the sense used in (narrow) information theory refers
neither to the process of telling something to somebody nor to the something
that is being told; indeed, it has nothing to do with meaning, not even in a
metaphoric sense. Instead, it is the statistical probability of a sign or signal
being selected from a given set of signs or signals. This probability is of
significance for the design of communication systems, for determining the
optimal size of a communication channel, for measuring its capacity for
signal transmission, for appraising the efficiency of a system (existing or
proposed), and for other purposes. Expressions such as “information in the
technical sense,” “statistical sense,” “mathematical sense,” or “engineer¬
ing sense,” have become customary in order to distinguish this concept from
25 As Donald MacKay writes in his paper in Section 7 of this volume, “[Shannon’s] theory did
not define the concept of information at all.”
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 659
A Compendium of Meanings
This second piece about semantic quirks should not take nearly as much
space as the previous one. Not that the word science has been used in fewer
662 FRITZ MACHLUP
diverse meanings than the word information; but the differences in the
meanings of science can be more easily explained.
Dictionary Definitions
The Latin noun scientia, derived from the verb scio, scire (to know), means
knowledge (both knowing-what and knowing-how), awareness, cognition,
insight, skill, and science in the widest sense. Some writers, Caesar among
them, spoke of scientia atque usus militum, which alludes to a possible
differentiation between theoretical knowledge and practical experience in
military matters, but a broad sense of “knowledge” was the more common
meaning of scientia.
The Oxford English Dictionary gives the following meanings: (1) The state
of fact of knowing; knowledge or cognizance. (2) Knowledge acquired by
study, b. Trained skill. (3) A particular branch of knowledge or study; a
recognized department of learning. (4) A branch of study which is concerned
either with a connected body of demonstrated truths or with observed facts
systematically classified . . . under general laws. (5) The kind of knowledge
or intellectual activity .... What is taught in the Schools, or may be learned
by study, b. In modern use often ‘Natural and Physical Science’, c. The
portions of philosophy, logic, etc., included in the course of study in the
Oxford School of Literae Humaniores.
Note the division of Entry (5) into (a) any discipline taught, (b) natural and
physical sciences, and (c) humanistic studies, where (b) is given as the
modern use and (c) as having become obsolete in the twentieth century. This
list reflects a history of semantic change but does not do justice to it: It is a
history of arbitrary restrictions imposed by successive schools of thought,
each trying to exclude the teachings of its opponents as unscientific. I have
elsewhere presented a brief sketch of this semantic development, but I think
that a still more concise statement here will be helpful. [Machlup, 1980, pp.
62-70.]
The sequence of exclusions has some similarities to a game of musical
chairs. For the playful the game may be good fun; in actual fact, however, it
has caused, and continues to cause, a waste of scientists’ time as each tries
to secure for himself a safe chair—instead of doing more important work.
The next stage in the semantic development was really one of liberalization,
in that the exclusion of empirical knowledge was lifted to the extent that
such knowledge was systematized according to stated principles. The Ger¬
man philosopher Immanuel Kant was among the leaders in making science
less exclusive; he still reserved the designation “science proper” for bodies
of knowledge formulated in propositions of apodictic certainty, that is,
purely analytic propositions. And metaphysics was still “the science of the
first principles of human cognition.” (Kant, 1848, p. 581.) But what had
previously been natural philosophy and moral philosophy became two cate¬
gories of empirical sciences, along with the formal sciences.
world. None of the equivalents of the word science in languages other than
English has that particular restrictive meaning. The academies of sciences
everywhere but in the English-speaking world retained their divisions for
historical sciences, juridical sciences, philosophical sciences, and so forth;
how can one translate these words into English without contradiction in
terms? (The phrase “arts and sciences” is often called in for the purpose.)
And how should “scientists” in the restrictive sense be translated into
French, German, Italian, Japanese, or Russian? (It can be done, of course,
by adding some additional words to modify the noun.) Far more serious than
the task of translation have been the countless confusions when partners in
learned discussion erroneously confounded the exclusive designation with
the wider meanings of science.26
26The title of Max Weber’s essay “Wissenschaft als Beruf” was rendered in English as “Sci¬
ence as a Vocation.” Since Weber meant by science all disciplines or branches of learning, the
translator should have rendered it as “academic disciplines” or “science and other scholar¬
ship,” or some other awkward phrase. Titles of editions of the scientific papers of eminent
lawyers, art historians, or theologians in Europe could not be translated into English before
checking whether the papers were in the authors’ own fields or perhaps on subjects of natural
science. After all, in Germany or France a lawyer’s paper on a legal issue prepared for publica¬
tion in a law journal is a scientific paper; if it were a brief for the court’s consideration, it would
be of direct practical usefulness and hence “nonscientific.” Erudition and purpose—not the
subject and not the method—determine what is science in cosmopolitan usage.
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 665
Excluding Philosophy
Some may think that the exclusion of philosophy is tantamount to the exclu¬
sion of nontestable knowledge. This is not so. The concept of science that
excludes philosophy may still include many hypotheses and propositions of
a hypothetico-deductive character as well as models containing ‘‘inter¬
mediate” variables that are nonmeasurable and nonobservable. A concept of
science that excludes nontested and nontestable knowledge-claims is, there¬
fore, more restrictive than the one that excludes philosophy.
[Editor’s Note (U.M.): This subsection was unfinished at the time of Professor Machlup’s
death. Also missing from the manuscript were the following subsections listed in his outline for
inclusion at this point: Excluding Nonempirical Knowledge; Excluding Nontheoretical, Non-
nomological Knowledge; Excluding Knowledge of Man and Society; Excluding Ideographic
and Historical Knowledge; Excluding Literary Scholarship in the Humanities; Excluding
Evaluative and Prescriptive Knowledge; and Excluding Nonquantitative Knowledge. The
manuscript continued with the subsection, Excluding Art.]
Excluding Art
27One example: “If computer programming is to become an important part of computer re¬
search and development, a transition of programming from an art to a disciplined science must
be effected.” (Bauer, Juncosa, and Perlis, 1959, pp. 121-122.)
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 667
“action.” (Bentham, 1816, p. 9.) Even with these categories there are
difficulties, in that the scientist too does something—experimenting, writ¬
ing, teaching—and the performer ordinarily thinks when using the special
knowledge needed for the attainment of a task.
Another distinction was proposed by John Stuart Mill: “Science is a
collection of truths; art, a body of rules, or directions for conduct.” (John
Stuart Mill, 1836 and 1844, p. 124.) This seems clear until we realize that
rules for conduct could easily be translated into general propositions of
cause and effect, and vice versa. As we look for criteria to distinguish
science and art, we will find it necessary to subdivide each. The distinction
between basic and applied science is unavoidable; as to art, we ought to
realize that (1) general teaching of rules for performance, (2) precepts for
concrete situations, (3) directives for immediate action, and (4) execution of
such directives are all different things. [Machlup, 1980, p. 92.] In this divi¬
sion of the spectrum, the line that separates applied science from art as a
body of general rules for possible performance is by no means hard and fast.
This is not necessarily a disadvantage; sometimes it is better not to impose
arbitrary demarcation lines where differentiation is a matter of judgment.
In his Turing Award lecture in 1974, Donald Knuth addressed himself to
the question of “Computer Programming as an Art.” He clearly disas¬
sociated himself from the opinion “that there is something undesirable about
an area of human activity that is classified as an ‘art’; it has to be a science
before it has any real stature.” (Knuth, 1974a, p. 667.) One of his theses was
that programming was also “an art form, in an aesthetic sense.” Program¬
ming “can be like composing poetry or music,” it can give aesthetic plea¬
sure to the programmer as well as to the user, it can be “a beautiful thing.”
(Ibid., pp. 670, 672.) Of course, programming is also an art in the technical
sense, in that it takes know-how and skill to write a good program for a task
to be executed by the computer. On the other hand, the required knowing-
how is based on a goodly measure of scientific knowing-that, and the pro¬
gram is designed to assist in finding other pieces of knowing-that. Even the
purest scientist needs to have various skills; to comprehend, to reason, to
see previously overlooked problems, to ask the right questions, to devise
relevant experiments, to interpret the findings intelligently, and so forth. In
other words, the scientific researcher needs a lot of know-how—and this is
art. Undoubtedly, the mixture of know-how, know-that, and know-what-to-
find-out is different for different tasks and different researchers; there may
be tasks for which routine skills are the chief ingredients. On the whole,
assessments of the share of art and the share of science in research are
questionable. To some extent, such assessments enter into job descriptions
and the choice of staff titles, for example, in the distinction between profes¬
sional and technical personnel; but I doubt that this is what people have in
mind when they proclaim that some activity is only art, not science.
Moreover, I cannot help, when I hear such a pronouncement, asking the
rude question “So what?” If programming, modeling, experimenting, fore-
668 FRITZ MACHLUP
casting, and similar activities are art, not science, what bearing does this
have on the activity so demoted? Perhaps the assessor has some dictum in
mind like “science is true” whereas art is not aiming at truth. The answer,
then, would be that the mentioned activities are engaged in the service of
finding, or approaching, the truth, even if no scientist can ever be sure to
have captured it.
Excluding Technology
28The concept of “practical science” goes back to Aristotle and held an important place in the
system of four “orders of science” distinguished by Thomas Aquinas—natural philosophy,
rational philosophy, moral philosophy, and practical sciences. Aquinas used “mechanical arts”
as a synonym for “practical sciences.” As examples he listed medicine, shipbuilding, and
strategy. [Machlup, 1982, pp. 26-27.]
29This statement seems to contradict the contention of Pearson and Slamecka in their paper to
the effect that “one of the distinguishing features of science and technology concerns the
essential role played in science by declarative sentences and in technology by imperative
sentences.” This point holds equally for a handbook of medicine: Handbooks and reference
books in medicine and in technology (computer technology, fibre technology, mining technol¬
ogy, and so forth) are not actually written in the imperative mood. Only their pragmatic applica¬
tions are closer to those in the foundation sciences, and to emphasize this, some philosophers
have resorted to the cookbook analogy.
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 669
Some people are very proud of the continuing changes of their language as
old words acquire new meanings. They take this as a sign of progress: Theirs
is a living language, which flexibly adapts itself to changes in the environ¬
ment, particularly in technology. They are not disturbed by the fact that
some of the new meanings of old words are largely due to ignorance of the
original meanings or to laziness, avoiding the task of coining new words that
30A subtle comment on the triple relationship among science, philosophy, and technology may
be quoted here: “The union between science and philosophy was possible only during a period
of separation between science and technology. Modem science was born when technology
became scientific. The union of science and technology was responsible for the separation
between science and philosophy.” (Philipp Frank, 1957, p. 29.)
31 In discussing the alleged juxtaposition of science, on the one hand, and practical arts, technol¬
ogy, and engineering, on the other, one may ask, as Saul Gorn does in his lead paper in this
volume, why the words academic and practical are so often used as mutual insults. The reason,
I submit, is that people often like to explain their career choices by appeals to higher principles:
The academic looks down on the practical man who is obtuse to the eternal values of the search
for learning for the sake of learning; the man of practice looks down on the long-haired “college
prof” who wastes his time on merely “academic” matters and is oblivious to practical affairs.
Both groups may be happy in their “superior wisdom.”
670 FRITZ MACHLUP
Dictionary Definitions
The word “to compute” is a direct derivation from the Latin computare,
where it has the same meaning as in English: “[T]o determine by computa¬
tion; to reckon, count; to take account of; [and] to make computation.” Its
synonym is “to calculate.”
The computing engine or computing machine, later called the computer,
was designed for number manipulation by means of symbol manipulation.
Charles Babbage (1792-1871), realizing that his machine could do more than
numerical calculations, called his device an analytical engine.33 The elec¬
tronic machines developed in the twentieth century, though capable of ma¬
nipulating symbols of any sort, were nevertheless designated as computers,
since the primary use for which they were intended was numerical computa¬
tion.
Multipurpose Machines
It did not take long for nonnumerical tasks to develop, to grow, and to
overtake the numerical ones. Some enterprising people might have come
forth with a new name for that miraculous machine that could do so many
things besides computing—but this did not happen or, at least, any new
names suggested were not accepted. So the noun “computer” stuck even
after it was generally known that the computer was capable of doing scores
of things other than computing. As a matter of fact, many computer experts
began to extend the meaning of the verb “computing” to make it comprise
virtually everything that can be done with a computer. This semantic expan¬
sion has caused some, though not serious, obstacles in professional and
scientific communication.
32 Among the stock examples of a long time ago are manufacture (meaning made by hand, now
used for the opposite, made by machine) and manuscript (meaning written by hand, now
including the opposite, a typescript).
33 “Meanwhile ... he [Babbage] was at work on . . . his Analytical Engine, a grander, bigger,
all-purpose calculating machine. It would not only be capable of arithmetic calculations, but it
would also be capable of analysis and of tabulating any function whatever.” (McCorduck, 1979,
pp. 25-26.)
SEMANTIC QUIRKS IN STUDIES OF INFORMATION 671
34When I requested Zenon Pylyshyn’s view on this semantic question, he replied: “Although
the term does have its roots in early calculating machines, my impression is that in psychology,
philosophy of mind, and certainly computer science it is viewed as an entirely neutral term
which designates what computers do—i.e., manipulate symbols. The public, of course, still
equates that with calculation over numbers, but in this case it is the public’s idea which will
change, rather than the term.” (Pylyshyn in private correspondence, August 20, 1982.)
CUMULATIVE LIST OF REFERENCES
673
674 CUMULATIVE LIST OF REFERENCES
Anderson, John R., Language, Memory, and Thought (Hillsdale, N. J.: Lawrence
Erlbaum, 1976).
Anderson, John R., and Bower, Gordon H., Human Associative Memory (Wash¬
ington, D.C.: V. H. Winston, 1973).
Aoki, Masanao, Optimal Control and Systems Theory in Dynamic Economic Analy¬
sis (New York: North-Holland, 1976).
Arbib, Michael A., “A Partial Survey of Cybernetics in Eastern Europe and the
Soviet Union,” in Behavioral Science, vol. 11 (1966), pp. 193-216.
Arbib, Michael A., The Metaphorical Brain: An Introduction to Cybernetics as
Artificial Intelligence and Brain Theory (New York: Wiley-Interscience, 1972).
Arbib, Michael A., “Artificial Intelligence and Brain Theory: Unities and Diver¬
sities,” in Annals of Biomedical Engineering, vol. 3 (Sept. 1975a), pp. 238-274.
Arbib, Michael A., “Cybernetics After 25 Years: A Personal View of System
Theory and Brain Theory,” in IEEE Transactions on Systems, Man, and Cyber¬
netics, vol. SMC-5 (1975b), pp. 359-363.
Arbib, Michael A., “Perceptual Structures and Distributed Motor Control,” in
Brooks, Vernon B., ed., Handbook of Physiology: Section 1: The Nervous
System, Vol. II: Motor Control (Bethesda, Md.: The American Physiological
Society, 1981), pp. 1449-1480.
Arbib, Michael A., “Modeling Neural Mechanisms of Visuomotor Coordination in
Frog and Toad,” in Amari, Shun-ichi, and Arbib, Michael A., eds., Competition
and Cooperation in Neural Nets (New York: Springer-Verlag, 1982), pp. 342-
370.
Arbib, Michael A.; Caplan, David; and Marshall, John C.,eds., Neural Models of
Language Processes (New York: Academic Press, 1982).
Arbib, Michael A., and Manes, Ernest G., “Machines in a Category,” in Journal of
Pure and Applied Algebra, vol. 19 (Dec. 1980), pp. 9-20.
Arden, Bruce W., ed., What Can Be Automated? The Computer Science and En¬
gineering Study (COSERS) (Cambridge, Mass.: MIT Press, 1980a).
Arden, Bruce W., “COSERS Overview,” in Arden, Bruce W., ed.. What Can Be
Automated? The Computer Science and Engineering Study (COSERS) (Cam¬
bridge, Mass.: MIT Press, 1980b), pp. 3-31.
Aristotle, The Nicomachean Ethics of Aristotle, with English notes by J. S. Brewer
(Oxford, England: Henry Slatter, 1836; another ed., New York: G. P. Putnam
Sons, 1926). [1st ed., 1513]
Armer, Paul, “Attitudes Toward Intelligent Machines,” in Feigenbaum, Edward
A., and Feldman, Julian, eds., Computers and Thought (New York: McGraw-
Hill, 1963), pp. 389-405.
Arnol’d, V. I., “Lectures on Bifurcations in Versal Families,” in Russian Mathe¬
matical Surveys, vol. 27 (Sept./Oct. 1972), pp. 54-123.
Arrow, Kenneth J., “Economic Equilibrium,” in Sills, David L., ed., International
Encyclopedia of the Social Sciences, vol. 4 (New York: Macmillan and Free
Press, 1968), pp. 376-388.
Arrow, Kenneth J., The Limits of Organization (New York: W. W. Norton, 1974).
Arrow, Kenneth J., “The Economics of Information,” in Dertouzos, Michael L.,
and Moses, Joel, eds., The Computer Age: A Twenty-Year View (Cambridge,
Mass.: MIT Press, 1979), pp. 306-317.
CUMULATIVE LIST OF REFERENCES 675
Ashby, W. Ross, Design for a Brain: The Origin of Adaptive Behaviour (New York:
Wiley, 1952; 2nd ed., rev., 1960).
Ashby, W. Ross, An Introduction to Cybernetics (London: Chapman and Hall, 1956).
Asratyan, G. E., and Simonov, P., How Reliable is the Brain? (Moscow: Mir, 1975).
Atkinson, Anthony A., and Feltham, Gerald A., “Information in Capital Markets:
An Agency Theory Perspective,” Working Paper, Faculty of Commerce and
Business Administration, University of British Columbia, Vancouver, B.C.,
1980.
Attneave, Fred, Applications of Information Theory to Psychology: A Summary of
Basic Concepts, Methods, and Results (New York: Holt, 1959).
Augustinus, Aurelius, St. Augustine’s Confessions (Cambridge, Mass.: Harvard
University Press, 1950-1951).
Backus, John, et al., “The FORTRAN Automatic Coding System,” in Proceedings
of the Western Joint Computer Conference, sponsored by the Institute of Radio
Engineers, the American Institute of Electrical Engineers, and the Association
for Computing Machinery and held [in Los Angeles] Feb. 26-28, 1957; pp. 188—
198.
Backus, John, “Can Programming Be Liberated From the von Neumann Style? A
Functional Style and Its Algebra of Programs” [the 1977 ACM Turing Award
Lecture], in Communications of the ACM, vol. 21 (Aug. 1978), pp. 613-641.
Baiman, Stanley, “Agency Research in Managerial Accounting: A Survey,” in
Journal of Accounting Literature, vol. 1 (spring 1980), pp. 154-213; reprinted in
Mattessich, Richard, et al., Modern Accounting Research: History, Survey, and
Guide (Vancouver, B.C.: Canadian Certified General Accountants Research
Foundation, forthcoming 1984).
Baker, Jeffrey J. W., and Allen, Garland E., The Study of Biology (Reading,
Mass.: Addison-Wesley, 4th ed., 1982).
Ballard, Dana H., and Brown, Christopher M., Computer Vision (Englewood
Cliffs, N.J.: Prentice-Hall, 1982).
Baralt-Torrijos, Jorge, “A Programmatic Interpretation of Combinatory Logic”
(Ph.D. diss., Georgia Institute of Technology, 1973).
Bar-Hillel, Yehoshua, Language and Information (Reading, Mass.: Addison-
Wesley, 1964).
Bar-Hillel, Yehoshua, and Carnap, Rudolf, “Semantic Information,” in Jackson
of Burnley, Willis Jackson, Baron, ed., Communication Theory. Papers read at a
symposium on Applications of Communication Theory, held [at the Institution
of Electrical Engineers, London] Sept. 22-26, 1952 (London: Butterworths,
1953a), pp. 503-512.
Bar-Hillel, Yehoshua, and Carnap, Rudolf, “Semantic Information,” in The Brit¬
ish Journal for the Philosophy of Science, vol. 4 (Aug. 19536), pp. 147-157.
Barker, Roger G., and Wright, Herbert F., The Midwest and Its Children: The
Psychological Ecology of an American Town (Evanston, Ill.: Row, Peterson,
1954; New York: Harper and Row, 1955).
Barnea, Amir; Haugen, Robert A.; and Senbet, Lemma W., “Market Imperfec¬
tions, Agency Problems, and Capital Structure: A Review,” in Financial Man¬
agement, vol. 10 (summer 1981), pp. 7-22.
Barr, Avron; Bennett, James S.; and Clancey, William J., Transfer of Expertise:
676 CUMULATIVE LIST OF REFERENCES
Bever, Thomas G., “Normal Acquisition Processes Explain the Critical Period for
Language Learning,” in Diller, Karl C., ed.. Individual Differences and Univer¬
sal in Language Learning Aptitude (Rowley, Mass.: Newbury House, 1980/?),
pp. 176-198.
Bever, Thomas G., and Hurtig, R. R., “Detection of a Non-Linguistic Stimulus is
Poorest at the End of a Clause,” in Journal of Psycholinguists Research, vol. 4
(1975), pp. 1-7.
Bhagavad-Gita, The, trans. Christopher Isherwood (London: Phoenix House, 1947).
Binet, Alfred, and Simon, Theodore, The Development of Intelligence in Children:
The Binet-Simon Scale, trans. Elisabeth S. Kite (Baltimore: Williams and Wil¬
kins, 1916).
Bitzer, Donald, “The Wide World of Computer-Based Education,” in Rubinoff,
Morris, and Yovits, Marshall C., eds., Advances in Computers, vol. 15 (New
York: Academic Press, 1976), pp. 239-284.
Black, Max, “The Definition of Scientific Method,” in Stauffer, Robert C., ed..
Science and Civilization (Madison: University of Wisconsin Press, 1949), pp.
67-95.
Blaivas, Alexander, “Visual Analysis,” in Mathematical Biosciences, vol. 28
(1977), pp. 101-149.
Bloomfield, Leonard, An Introduction to the Study of Language (New York: Holt,
1914).
Bloomfield, Leonard, Language (New York: Holt, 1933).
Bloomfield, Leonard, “Linguistic Aspects of Science,” in International Encyclope¬
dia of Unified Science, vol. 1, no. 4 (Chicago: University of Chicago Press,
1939).
Blumenthal, Arthur L., The Process of Cognition (Englewood Cliffs, N.J.: Pren¬
tice-Hall, 1977).
Bobrow, Daniel G., “Dimensions of Representation,” in Bobrow, Daniel G., and
Collins, Allan, eds., Representation and Understanding: Studies in Cognitive
Science (New York: Academic Press, 1975), pp. 1-34.
Bobrow, Daniel G., ed., Special Issue on Non-Monotonic Logic, Artificial Intelli¬
gence, vol. 13 (Apr. 1980), pp. 1-172.
Bobrow, Daniel G., and Collins, Allan, eds,.. Representation and Understanding:
Studies in Cognitive Science (New York: Academic Press, 1975).
Bobrow, Daniel G., and Raphael, Bertram, “New Programming Languages for
Artificial Intelligence Research,” in ACM Computing Surveys, vol. 6 (Sept.
1974), pp. 153-174.
Bobrow, Robert J., and Brown, John Seely, “System Understanding: Synthesis,
Analysis, and Contingent Knowledge in Specialized Understanding Systems,”
in Bobrow, Daniel G., and Collins, Allan, eds.. Representation and Understand¬
ing: Studies in Cognitive Science (New York: Academic Press, 1975), pp. 103—
129.
Bode, Hendrik; Mosteller, Frederick; Tukey, John; and Winsor, Charles, “The
Education of a Scientific Generalist,” in Science, vol. 109 (June 1949), pp. 553—
558.
Boden, Margaret A., Purposive Explanation in Psychology (Cambridge, Mass.:
Harvard University Press, 1972).
CUMULATIVE LIST OF REFERENCES 679
Boden, Margaret A., Artificial Intelligence and Natural Man (New York: Basic
Books, 1977).
Boden, Margaret A., Minds and Mechanisms: Philosophical Psychology and Com¬
putational Models (Ithaca, N.Y.: Cornell University Press, 1981a).
Boden, Margaret A., “The Case for a Cognitive Biology,” in Boden, Margaret A.,
Minds and Mechanisms: Philosophical Psychology and Computational Models
(Ithaca, N.Y.: Cornell University Press, 19816), pp. 89-112.
Bogdanov [Malinovskii], Aleksandr Aleksandrovich, Vseobshchaia organization-
naia nauka, tektologiia [Tektology: The Universal Science of Organization],
3 vols. (St. Petersburg: Izdatelstvo, 1913; 2nd ed., Moscow: Izdatelstvo,
1922).
Boulding, Kenneth E., The Image: Knowledge in Life and Society (Ann Arbor:
University of Michigan Press, 1956a).
Boulding, Kenneth E., “General Systems Theory—The Skeleton of Science,” in
Management Science, vol. 2 (Apr. 1956b), pp. 197-208.
Boulding, Kenneth E., The Meaning of the Twentieth Century (New York: Harper,
1964).
Boulding, Kenneth E., “The Economics of Knowledge and the Knowledge of
Economics,” in American Economic Review, vol. 56, no. 2 (1966), pp. 1-13.
Boulding, Kenneth E., Eco-dynamics: A New Theory of Societal Evolution (Bev¬
erly Hills, Calif.: Sage Publications, 1978).
Bowler, T. Downing, General Systems Thinking: Its Scope and Applicability (New
York: Elsevier/North-Holland, 1981).
Brachman, Ronald J., and Smith, Brian C., eds., Special Issue on Knowledge
Representation, SIGART Newsletter, vol. 70 (Feb. 1980), pp. 1-138.
Branscomb, Lewis M., “Information: The Ultimate Frontier,” in Science, vol. 203
(12 Jan. 1979), p. 143.
Bresnan, Joan, “A Realistic Transformational Grammar,” in Halle, Morris; Bres-
nan, Joan; and Miller, George A., eds., Linguistic Theory and Psychological
Reality (Cambridge, Mass.: MIT Press, 1978), pp. 1-59.
Briet, Suzanne, Qu’est-ce que la documentation? (Paris: Editions Documentaires
Industrielles et Techniques, 1951).
Brillouin, Leon, “Negentropy Principle of Information,” in Journal of Applied
Physics, vol. 24 (Sept. 1953), pp. 1152-1163.
Brillouin, Leon, Science and Information Theory (New York: Academic Press,
1956; 2nd ed. 1962).
Brockett, Roger W., Finite Dimensional Linear Systems (New York: Wiley, 1971).
Brockett, Roger W., “The Geometry of Partial Realization Problems,” in Proceed¬
ings of a Conference on Decision and Control organized by the Institute of
Electrical and Electronics Engineers (New York: IEEE Press, 1978).
Broglie, Louis de, “L’Avenir de la physique,” in Broglie, Louis de, et al., L’Avenir
de la science (Paris: Plon, 1941), pp. 1-35.
Broglie, Louis de, La cybernetique: theorie du signal et de Tinformation (Paris:
Editions de la Revue d’Optique Theorique et Instrumentale, 1951), p. v.
Brookes, Bertram C., “Robert Fairthorne and the Scope of Information Science,”
in Journal of Documentation, vol. 30 (June 1974), pp. 139-152.
Brookes, Bertram C., “A New Paradigm for Information Science,” in The Informa¬
tion Scientist, vol. 10 (Sept. 1976), pp. 103-111.
Brown, Robert R., Explanation in Social Science (Chicago: Aldine-Atherton, 1963).
Brunovsky, P., “A Classification of Linear Controllable Systems,” in Kybernetika,
vol. 6 (1970), pp. 176-188.
Bryson, Arthur, and Ho, Y. C., Applied Optimal Control (New York: Halsted
Press, 1968).
Buchanan, James, “The Justice of Natural Liberty,” in O’Driscoll, Gerald P., Jr.,
ed., Adam Smith and Modern Political Economy: Bicentennial Essays on The
Wealth of Nations (Ames: Iowa State University Press, 1979) pp. 117-131.
CUMULATIVE LIST OF REFERENCES 681
Buckland, Michael K., Library Services in Theory and Context (New York: Perga-
mon Press, 1983).
Buckley, Walter, Sociology and Modern Systems Theory (Englewood Cliffs, N.J.:
Prentice-Hall, 1967).
Bunge, Mario, Treatise on Basic Philosophy, Vol. IV, Ontology II: A World of
Systems (Dordrecht, Holland, and Boston: D. Reidel, 1979a).
Bunge, Mario, “A Systems Concept of Society: Beyond Individualism and Hol¬
ism,” in Theory and Decision, vol. 10 (Jan. 1979b), PP- 13-30.
Burks, Arthur W., ed., Essays on Cellular Automata (Urbana: University of Illnois
Press, 1970).
Burmeister, Edwin, and Dobell,A. Rodney, Mathematical Theories of Economic
Growth (New York: Macmillan, 1970).
Bush, Vannevar, “As We May Think” (1945). Reprinted in Pylyshyn, Zenon W.,
ed., Perspectives on the Computer Revolution (Englewood Cliffs, N.J.: Pren¬
tice-Hall, 1970), pp. 47-59.
Butler, Pierce, An Introduction to Library Science (Chicago: University of Chicago
Press, 1933).
Butterfield, Herbert,The Origins of Modern Science 1300-1800 (London: Bell,
1949; New York: Macmillan, 1951; Toronto: Clark, Irwin and Co., 1957).
Byrnes,C., and Falb, Peter L. “Applications of Algebraic Geometry in System
Theory,” in American Journal of Mathematics, vol. 101 (1979), pp. 337-363.
Cadwallader, M., “The Cybernetic Analysis of Change in Complex Social Organi¬
zations,” in American Journal of Sociology, vol. 65 (1959), pp. 154-157.
Cannon, Walter B., “Organization for Physiological Homeostasis,” in Physiologi¬
cal Reviews, vol. 9 (July 1929), pp. 399-431.
Cannon, Walter B., The Wisdom of the Body (New York: W. W. Norton, 1939).
Capek, Karel, R.U.R. (Rossum’s Universal Robots) (Garden City, N.Y.: Double¬
day, Page and Co., 1923).
Carnap, Rudolf, “Foundations of Logic and Mathematics,” in Neurath, Otto; Car¬
nap, Rudolf; and Morris, Charles, eds., Foundations of the Unity of Science:
Toward an International Encyclopedia of Unified Science, vol. 1 (Chicago: Uni¬
versity of Chicago Press, 1939a; bound ed., 1969), pp. 139-214.
Carnap, Rudolf, “Logical Foundations of the Unity of Science,” in Neurath, Otto;
Carnap, Rudolf; and Morris, Charles, eds., Foundations of the Unity of
Science: Toward an International Encyclopedia of Unified Science, vol. 1
(Chicago: University of Chicago Press, 1939b; bound ed., 1969).
Carnap, Rudolf, Logical Foundations of Probability (Chicago: University of
Chicago Press, 1950).
Carnap, Rudolf, and Bar-Hillel, Yehoshua, An Outline of a Theory of Semantic
Information (Cambridge, Mass.: MIT Research Laboratory of Electronics Tech¬
nical Report, 1952). Reprinted as chap. 15 in Bar-Hillel, Language and Informa¬
tion: Selected Essays on Their Theory and Application (Reading, Mass.: Addi-
son-Wesley, 1964).
Carroll,C. Edward, The Professionalization of Education for Librarianship
(Metuchen, N.J.: Scarecrow Press, 1970).
682 CUMULATIVE LIST OF REFERENCES
Collins, Allan, “Why Cognitive Science?” in Cognitive Science, vol. 1 (Jan. 1977),
pp. 1-2.
Converse, Alvin O., “The Use of Uncertainty in a Simultaneous Search,” in Opera¬
tions Research, vol. 15 (Nov./Dec. 1967), pp. 1088-1095.
Copinger, Walter Arthur, “On the Necessity for the Formation of a Biblio¬
graphical Society of the United Kingdom and Suggestions as to its Operations,”
in The Library, vol. 4 (Jan./Dec. 1892), pp. 1-7.
Costa, Serpil de, “Foundation and Development of IFLA, 1926-1939,” in The
Library Quarterly, vol. 52 (Jan. 1982), pp. 41-58.
Costanza, Robert, “Embodied Energy and Economic Valuation,” in Science, vol.
210 (12 Dec., 1980), p. 1224.
Counelis, James S., “Cybernetics and General Systems—A Unitary Science (Infor¬
mation and the Unity of General Systems Theory and Cybernetics),” in Kyber-
netes, vol. 8 (1979), pp. 25-32.
Cowan, W. Maxwell, ed., Annual Review of Neuroscience (Palo Alto, Calif.: An¬
nual Reviews Inc., 1981).
Craik, Kenneth J. W., The Nature of Explanation (Cambridge: Cambridge Univer¬
sity Press, 1943).
Croce, Benedetto, “The Task of Logic,” in Encyclopaedia of the Philosophical
Sciences, Vol. 1: Logic (London: Macmillan, 1913).
Cuatrecasas, Pedro, and Greaves, Melvyn F., eds., Receptors and Recognition,
series A, vols. 1, 2 (London: Chapman and Hall, 1976).
Cyert, Richard M., and March, James G., A Behavioral Theory of the Firm (En¬
glewood Cliffs, N. J.: Prentice-Hall, 1963).
Cyranski, John F., “Measurement, Theory, and Information,” in Information and
Control, vol. 41 (June 1979), pp. 275-304.
Davis, Martin D., ed., The Undecideable: Basic Papers on Undecideable Proposi¬
tions, Unsolvable Problems, and Uncomputable Functions (Hewlett, N. Y.:
Raven Press, 1965).
Davis, Randall, “Applications of Meta-Level Knowledge to the Construction,
Maintenance, and Use of Large Knowledge Bases,” in Davis, Randall, and
Lenat, Douglas, Knowledge-Based Systems in Artificial Intelligence (New
York: McGraw-Hill, 1982), pp. 229-390.
Davis, Randall, and Lenat, Douglas, Knowledge-Based Systems in Artificial Intel¬
ligence (New York: McGraw-Hill, 1982).
Davis, Watson, “Regarding the Name: Documentation Institute of Science Service”
(Washington, D.C.: Documentation Institute of Science Service, 11 July 1935),
mimeographed.
Dawkins, Richard, The Selfish Gene (New York and London: Oxford University
Press, 1976).
Day, Richard H., “A Microeconomic Model of Business Growth, Decay and Cy¬
cles,” in Unternehmensfusching Band 11, Heft 1 (1967), pp. 1-20.
Day, Richard H., “Adaptive Processes and Economic Theory,” in Day, Richard H.,
and Groves, Theodore, eds., Adaptive Economic Models, proceedings of a Sym¬
posium on Adaptive Economics, held [at the University of Wisconsin, Madison]
Oct. 21-23, 1974 (New York: Academic Press, 1975).
Day, Richard H., “Dynamic Systems and Epochal Change,” in Sabloff, Jeremy A.,
CUMULATIVE LIST OF REFERENCES 685
Ermentrout, G. B., and Cowan, Jack D., “Large-Scale Spatially Organized Activ¬
ity in Neural Nets,” in SIAM Journal on Applied Mathematics, vol. 38 (Feb.
1980), pp. 1-21.
Ernst, George W., and Newell, Allen, GPS: A Case Study in Generality and
Problem Solving (New York: Academic Press, 1969).
Evans, Robert B., “Basic Relationships Among Entropy, Exergy, Energy, and
Availability,” in Spiegler, K. S., ed., Principles of Desalination (New York:
Academic Press, 1966), pp. 44-66.
Evans, Thomas G., “A Program for the Solution of Geometric-Analogy Intelligence
Test Questions,” in Minsky, Marvin L., ed., Semantic Information Processing
(Cambridge, Mass.: MIT Press, 1968), pp. 271-354.
Ewert, Jorg-Peter, Neuroethology: An Introduction to the Neurophysiological
Fundamentals of Behavior (Berlin and New York: Springer-Verlag, 1980).
Falb, Peter L., “Linear Systems and Invariants” (Lecture Notes, Control Group,
Lund University, Sweden, 1974).
Fama, Eugene F., “Agency Problems and the Theory of the Firm,” in Journal of
Political Economy, vol. 88 (Apr. 1980), pp. 288-307.
Fano, Robert M., Transmission of Information (Cambridge, Mass., and New York:
MIT Press and Wiley, 1961).
Farradane, Jason E. L., “The Institute: The First Twelve Years,” in The Informa¬
tion Scientist, vol. 4 (Dec. 1970), pp. 143-151.
Faurre, P., “Realisations markoviennes de processus stationnaires,” Research Re¬
port No. 13, INRIA (Rocquencourt, France: Institut National de Recherche en
Informatique et en Automatique, 1973).
Faurre, P.; Clerget, M.; and Germain, F., Operateurs rationnels positifs: Applica¬
tion a Vhyperstabilite et aux processus aleatoires (Paris: Dunod, 1979).
Feigenbaum, Edward A., “The Simulation of Verbal Learning Behavior,” in Ex¬
tending Man’s Intellect, proceedings of the Western Joint Computer Confer¬
ence, held [in Los Angeles] by the Institute of Radio Engineers, the American
Institute of Electrical Engineers, and the Association for Computing Machinery,
May 9-11, 1961, vol. 19, pp. 121-132.
Feigenbaum, Edward A., “Artificial Intelligence: Themes in the Second Decade,” in
Morrell, A. J. H., ed., Information Processing 68, proceedings of the IFIP
Congress 1968, organized by the International Federation for Information Pro¬
cessing [Edinburgh, Scotland] Aug. 5-10, 1968 (Amsterdam: North-Holland,
1969), pp. 1008-1024.
Feigenbaum, Edward A., “The Art of Artificial Intelligence, I: Themes and Case
Studies of Knowledge Engineering,” in IJCAI-77: Proceedings of the Fifth In¬
ternational Joint Conference on Artificial Intelligence, held [at the Massachu¬
setts Institute of Technology, Cambridge, Mass.] by the International Joint Con¬
ference on Artificial Intelligence, Aug. 22-25, 1977 (Pittsburgh: Carnegie-Mellon
University), vol. 2, pp. 1014-1029.
Feigenbaum, Edward A., and Feldman, Julian, eds., Computers and Thought
(New York: McGraw-Hill, 1963).
Feigenbaum, Mitchell J., “Universal Behavior in Nonlinear Systems,” in Los
Alamos Science, vol. 1 (summer 1980), pp. 4-27.
CUMULATIVE LIST OF REFERENCES 689
Frank, Philipp, Philosophy of Science: The Link between Science and Philosophy
(Englewood Cliffs, N.J.: Prentice-Hall, 1957).
Frederiks, J. A. M., “Disorders of the Body Schema,” in Vinken, P. J., and Bruyn,
G. W., eds., Handbook of Clinical Neurology, Disorders of Speech, Perception,
and Symbolic Behavior (Amsterdam: North-Holland, 1969-), vol. 4, pp. 207-
240.
Frege, Friedrich Ludwig Gottlob, Begriffsschrift, eine der arithmetischen nach-
gebildete Formelsprache des reinen Denkens (Halle A/S, Germany: L. Nebert,
1879).
Freidin, Robert A., “Cyclicity and the Theory of Grammar,” in Linguistic Inquiry,
vol. 9 (1978), pp. 519-550.
Fried, Charles, Right and Wrong (Cambridge, Mass.: Harvard University Press,
1978) .
Fromkin, Victoria A., et al., Speech Errors as Linguistic Evidence (The Hague,
Netherlands: Mouton, 1973).
Fuhrman, P., “Functional Models, Factorizations, and Linear Systems”—talks
given at the NATO-AMS Advanced Study Institute on Algebraic and Geometric
Methods in Linear System Theory (Cambridge, Mass.: Harvard University,
1979) .
Furth, Hans G., “The Operative and Figurative Aspects of Knowledge and Infor¬
mation,” in Debons, Anthony, ed., Information Science: Search for Identity,
proceedings of the 1972 NATO Advanced Study Institute in Information Sci¬
ence, held [in Seven Springs, Champion, Penn.] Aug. 12-20, 1972 (New York:
Marcel Dekker, 1974), pp. 21-27.
Gabor, Dennis, “Theory of Communication,” in Journal of the Institution of Electri¬
cal Engineers, vol. 93, no. 3 (1946), pp. 429-457.
Gallagher, Robert G., Information Theory and Reliable Communication (New
York: Wiley, 1968).
Galvin, Thomas J., “The Challenge of Technology in the Eighties,” lecture pre¬
sented at Ann Arbor, University of Michigan, “Issues and Challenges of the
Eighties: A Michigan Libraries Forum,” 1981.
Gardner, Beatrice T., and Gardner, R. Allen, “Two-Way Communication with
an Infant Chimpanzee,” in Schrier, Allan M., and Stollnitz, Fred, edsBehavior
of Non-Human Primates: Modern Research Trends, vol. 4 (New York: Aca¬
demic Press, 1971), pp. 117-184.
Garfield, Eugene; Malin, Morton V.; and Small, Henry, “Citation Data as Sci¬
ence Indicators,” in Elkana, Yehuda, et al., eds., Toward a Metric of Science:
The Advent of Science Indicators (New York: Wiley, 1978).
Garner, Wendell R., Uncertainty and Structure as Psychological Concepts (New
York: Wiley, 1962).
Gatlin, Lila L., Information Theory and the Living System (New York: Columbia
University Press, 1972).
Gauld, Alan, and Shotter, John, Human Action and Its Psychological Investiga¬
tion (London and Boston: Routledge and Kegan Paul, 1977).
Gazis, Denos C.; Herman, Robert; and Potts, Renfrey B., “Car Following Theory
of Steady-State Traffic Flow,” in Operations Research, vol. 7 (July/Aug. 1959),
pp. 499-505.
CUMULATIVE LIST OF REFERENCES 691
Gorn, Saul, “The Identification of the Computer and Information Sciences: Their
Fundamental Semiotic Concepts and Relationships,” in Foundations of Lan¬
guage: International Journal of Language and Philosophy, vol. 4 (Nov. 1968),
pp. 339-372.
Greaves, Melvyn F., “Cell Surface Receptors: A Biological Perspective,” in Cuat-
recasas, Pedro, and Greaves, Melvyn F., eds., Receptors and Recognition, vol.
1 (London: Chapman and Hall, 1976), series A, pp. 1-32.
Gregory, Richard L., “On How So Little Information Controls So Much Behav¬
ior,” in Waddington, Conrad H., ed., Towards a Theoretical Biology, Vol. 2:
Sketches, proceedings of an IUBS Symposium (Chicago: Aldine-Atherton,
1969), pp. 236-247.
Greibach, Sheila A., “Formal Languages: Origins and Directions,” in Symposium
on the Foundations of Computer Science, proceedings of the Twentieth Annual
Symposium on Foundations of Computer Science, held [in San Juan, Puerto
Rico] by the IEEE Computer Society’s Technical Committee on Mathematical
Foundations of Computing (New York: Institute of Electrical and Electronics
Engineers, 1979), pp. 66-90.
Grenander, Ulf, Lectures in Pattern Theory, Vol. 1: Pattern Synthesis (New York:
Springer-Verlag, 1976).
Grenander, Ulf, Lectures in Pattern Theory, Vol. 2: Pattern Analysis (New York:
Springer-Verlag, 1978).
Groot, Adriaan D. de, Thought and Choice in Chess (The Hague, Netherlands:
Mouton, 1965).
Grossberg, Stephen, “Adaptive Pattern Classification and Universal Recoding: I.
Parallel Development and Coding of Neural Feature Detectors,” in Biological
Cybernetics, vol. 23, no. 2 (1976), pp. 121-134.
Gunderson, Keith, Mentality and Machines (Garden City, N.Y.: Doubleday, 1971).
Habermas, Jurgen, Toward a Rational Society: Student Protest, Science, and Poli¬
tics, trans. Jeremy J. Shapiro (Boston: Beacon Press, 1970).
Haberstroh, Chadwick J., “Control as an Organizational Process,” in Management
Science, vol. 6 (Jan. 1960), pp. 165-171.
Haken, Hermann, Synergetics: An Introduction: Nonequilibrium Phase Transitions
and Self-Organization in Physics, Chemistry, and Biology (Berlin and New
York: Springer-Verlag, 1977; 2nd ed., 1978).
Haken, Hermann, “Synergetics: Are Cooperative Phenomena Governed by Uni¬
versal Principles?” in Naturwissenschaften, vol. 67 (Mar. 1980), pp. 121-128.
Hall, Thomas Steele, Ideas of Life and Matter: Studies in the History of General
Physiology 600BC-1900AD, Vol. I: From Pre-Socratic Times to the Enlighten¬
ment (Chicago: University of Chicago Press, 1969).
Hanson, Allen R., and Riseman, Edward M., eds., Computer Vision Systems,
proceedings of a Workshop on Computer Vision Systems, held [at the Univer¬
sity of Massachusetts, Amherst] 1977 (New York: Academic Press, 1978).
Hardin, Garrett, Nature and Man’s Fate (New York: Mentor, 1961).
Hardin, Garrett, “The Tragedy of the Commons,” in Science, vol. 162 (1968), p.
1243.
Hardin, Garrett, Exploring New Ethics for Survival (New York: Viking Press,
1977).
CUMULATIVE LIST OF REFERENCES 693
book of Sensory Physiology, Vol. VIII: Perception (Berlin, Heidelberg, and New
York: Springer-Verlag, 1978).
Hempel, Carl G., “Fundamentals of Concept Formation in Empirical Science,” in
Neurath, Otto; Carnap, Rudolf; and Morris, Charles, eds., Foundations of the
Unity of Science: Toward an International Encyclopedia of Unified Science,
vol. 2 (Chicago: University of Chicago Press, 1939; bound ed., 1969), pp. 651 —
746.
Hempel, Carl G., Philosophy of Natural Science (Chicago: Aldine-Atherton, 1963).
Herman, E. Edward, and Kuhn, Alfred, Collective Bargaining and Labor Rela¬
tions (Englewood Cliffs, N.J.: Prentice-Hall, 1981).
Herman, Robert, and Montroll, Elliott W., “Some Statistical Observations from
a 75 Year Run of Sears Roebuck Catalogues,” in Landman, Uzi, ed., Statistical
Mechanics and Statistical Methods in Theory and Application: A Tribute to
Elliott W. Montroll, proceedings of a Symposium on Statistical Mechanics and
Statistical Methods in Theory and Research, held [at the University of Roches¬
ter, N.Y.] Nov. 4, 1976 (New York: Plenum Press, 1977), pp. 785-803.
Herman, Robert;Montroll, Elliott W.; Potts, Renfrey B.; and Rothery,
Richard W., “Traffic Dynamics: Analysis of Stability in Car Following,” in
Operations Research, vol. 7 (Jan./Feb. 1959), pp. 86-106.
Hermann, Robert T., and Martin, Clyde F., “Applications of Algebraic Geometry
to System Theory, Part I,” in IEEE Transactions on Automatic Control, vol.
AC-22 (Feb. 1977), pp. 19-25.
Hesse, Hermann, Magister Ludi (Das Glasperlenspeit) [the Nobel Prize novel],
trans. Mervyn Savill (New York: Holt, 1949).
Hick, W. Edmund, “On the Rate of Gain of Information,” in Quarterly Journal of
Experimental Psychology, vol. 4 (Feb. 1952), pp. 11-26.
Hilgard, Ernest R., “Pain Perception in Man,” in Held, Richard; Leibowitz,
Herschel W.; and Teuber, Hans-Lukas, eds., Handbook of Sensory Physiology,
Vol. VIII: Perception (Berlin, Heidelberg, and New York: Springer-Verlag,
1978), pp. 849-875.
Hilgard, Ernest R., and Bower, Gordon H., Theories of Learning (New York:
Appleton-Century-Crofts, 1948; 4th ed., 1975).
Hillman, Donald J., “Knowledge Transfer Systems,” in Weiss, Edward C., ed.,
The Many Faces of Information Science (Boulder, Colo.: Westview Press,
1977), pp. 75-103.
Hillman, James, The Myth of Analysis: Three Essays in Archetypal Psychology
(Evanston, Ill.: Northwestern University Press, 1972).
Hilton, Ronald W., “The Determinants of Information Value: Synthesizing Some
General Results,” in Management Science, vol. 27 (Jan. 1981), pp. 57-64.
Hinton, G. E., “Shape Representation in Parallel Systems,” in Proceedings of the
Seventh International Joint Conference on Artificial Intelligence [Vancouver,
B.C., 1981], pp. 1088-1096.
Hirshleifer, Jack, and Riley, JohnG., “The Analytics of Uncertainty and Infor¬
mation—An Expository Survey,” in Journal of Economic Literature, vol. 17
(Dec. 1979), pp. 1375-1421.
Hobbes, Thomas, Leviathan, or, The Matter, Form and Power of a Commonwealth
Ecclesiastical and Civil (London: Andrew Cooke, 1651).
CUMULATIVE LIST OF REFERENCES 695
Hofstadter, Douglas R., Godel, Escher, Bach: an Eternal Golden Braid (New
York: Basic Books, 1979).
Hofstadter, Douglas R., “Metamagical Themas: How Might Analogy, the Core of
Human Thinking, be Understood by Computers?,” in Scientific American,
Sept. 1981, p. 18.
Hofstadter, Douglas R., Who Shoves Whom Around Inside the Careenium?
(Bloomington: Indiana University Computer Science Department Technical Re¬
port No. 130, July 1982a; forthcoming in Synthese).
Hofstadter, Douglas R., ‘‘Metamagical Themas: Can Inspiration Be Mechanized?”
in Scientific American, Sept. 1982ft, p. 16.
Hofstadter, Douglas R., “Metamagical Themas: On Variations on a Theme as the
Essence of Imagination,” in Scientific American, Oct. 1982c, p. 20.
Hofstadter, Douglas R., The Tumult of Inner Voices (forthcoming, Southern Utah
State College, 1982d).
Hofstadter, Douglas R.; Clossman, Gray A.; and Meredith, Marsha J., Shake¬
speare’s Plays Weren’t Written by Him, But by Someone Else of the Same
Name: An Essay on Intensionality and Frame-Based Knowledge Representa¬
tion Systems (Bloomington: Indiana University Computer Science Department
Technical Report No. 96, July 1980).
Hofstadter, Douglas R., and Dennett, Daniel C., eds., The Mind’s I: Fantasies
and Reflections of Self and Soul (New York: Basic Books, 1981).
Holstrom, Bengt R., “Moral Hazard and Observability,” in Bell Journal of Eco¬
nomics, vol. 10 (spring 1979), pp. 74-91.
Holzner, Burkart, and Marx, John H., Knowledge Application: The Knowledge
System in Society (Boston: Allyn and Bacon, 1979).
Homans, George C., The Nature of Social Change (New York: Harcourt, Brace, and
World, 1967).
Hoos, Ida R., Systems Analysis in Public Policy: A Critique (Berkeley: University of
California Press, 1972).
Hopcroft, John E., and Ullman, Jeffrey D., Introduction to Automata Theory,
Languages, and Computation (Reading, Mass.: Addison-Wesley, 1979).
Hopf, Eberhard, “Abzweigung einer periodischen Losung von einer Stationaren
Losung eines Differentzialsystems,” in Berichte iiber die Verhandlungen Sach-
sischen Akademie der Wissenschaften zu Leipzig, Mathematisch-Naturwissen-
schaftliche Klasse, vol. 95 (1943), pp. 3-22.
Howard, Ronald A., “Information Value Theory,” in IEEE Transactions on Sys¬
tems Science and Cybernetics, vol. SSC-2 (Aug. 1966), pp. 22-26.
Hume, David, Dialogues Concerning Natural Religion (1779), in Green, T. H., and
Grose, T. H., eds., A Treatise on Human Nature (London: Longmans, Green,
and Coy., new ed., 1878, vol. II), pp. 380-468; reprinted in The Empiricists
(Garden City, N.Y.: Doubleday, 1961).
I Ching, or Book of Changes, trans. Richard Wilhelm and Cary F. Baynes (Prince¬
ton, N.J.: Princeton University Press, Bollingen Series 19, 1951).
IEEE Transactions on Information Theory (New York: Institute of Electrical and
Electronics Engineers, 1954, 1956).
Ingarden,R., and Urbanik, K., “Information Without Probability,” in Colloquium
Mathematicum, vol. 9 (1962), pp. 131-149.
696 CUMULATIVE LIST OF REFERENCES
Jewett, Charles C., “A Plan for Stereotyping Catalogues by Separate Titles; and for
Forming a General Stereotyped Catalogue of Public Libraries in the United
States,” in Proceedings of the American Association for the Advancement of
Science, vol. 4 (1851), pp. 165-176.
Johns, Ada Winifred, Special Libraries: Development of the Concept, their Organi¬
zations, and their Services (Metuchen, N.J.: Scarecrow Press, 1968).
Josiah Macy Jr. Foundation, Cybernetics, proceedings of the Sixth Conference
(1949) to the Tenth Conference (1953) (New York: Josiah Macy Jr. Foundation,
last pub. 1955).
Julesz, Bela, Foundation of Cyclopean Perception (Chicago: University of Chicago
Press, 1971).
Kalman, Rudolf E., “A New Approach to Linear Filtering and Prediction Prob¬
lems,” in Transactions of the American Society of Mechanical Engineers, Jour¬
nal of Basic Engineering, series D, vol. 82 D (Mar. 1960a), pp. 35-45.
Kalman, Rudolf E., “On the General Theory of Control Systems,” in Proceedings
of the First IFAC Congress, held [in Moscow] 1960 (London: Butterworths,
19606), vol. 1, pp. 481-491.
Kalman, Rudolf E., “Canonical Structure of Linear Dynamical Systems,” in Pro¬
ceedings of the National Academy of Sciences (USA), vol. 48 (1962), pp. 596-
600.
Kalman, Rudolf E., “Linear Stochastic Filtering Theory—Reappraisal and Out¬
look,” in Fox, J., ed., Proceedings of a Symposium on System Theory (Brook¬
lyn, N.Y.: Polytechnic Institute of Brooklyn, 1965), pp. 197-205.
Kalman, Rudolf E., “On Minimal Partial Realizations of a Linear Input/Output
Map,” in Kalman, Rudolf E., and De Claris, N., eds., Aspects of Network and
System Theory (New York: Holt, Rinehart, and Winston, 1971), pp. 385-408.
Kalman, Rudolf E., “Kronecker Invariants and Feedback,” in Weiss, L., ed.,
Proceedings of the 1971 NRL-MRC Conference on Ordinary Differential Equa¬
tions (New York: Academic Press, 1972), pp. 459-471.
Kalman, Rudolf E., “Comments on the Scientific Aspects of Modeling,” in Marois,
M., ed., Towards a Plan of Action for Mankind (Amsterdam: North-Holland,
1974), pp. 492-505.
Kalman, Rudolf E., “Realization Theory of Linear Dynamical Systems,” in Con¬
trol Theory and Functional Analysis (Vienna: International Atomic Energy
Agency, 1976).
Kalman, Rudolf E., “On Partial Realizations, Transfer Functions, and Canonical
Forms,” in Acta Polytechnica Scandinavica, Mathematics and Computer Sci¬
ence Series no. 31 (1979), pp. 9-32.
Kalman, Rudolf E., “A System-Theoretic Critique of Dynamic Economic Mod¬
els,” in International Journal of Policy Analysis and Information Systems, vol.
4 (1980), pp. 3-22.
Kalman, Rudolf E., “Identifiability and Problems of Model Selection in Econo¬
metrics”—invited paper presented at the Fourth World Congress of the Econ¬
ometric Society, held [in Aix-en-Provence, France] 1980 (rev. version, Jan.
1981).
Kalman, Rudolf E., “Finitary Models for Time Series and the Identification of
698 CUMULATIVE LIST OF REFERENCES
King, Donald W.; Krauser, Cheri; and Sague, Virginia M., “Profile of ASIS
Membership,” in Bulletin of the American Society for Information Science, vol.
6 (Aug. 1980), pp. 9-17.
King, Gilbert W., et al., Automation and the Library of Congress (Washington,
D.C.: Library of Congress, 1963).
Kirzner, Israel M., Competition and Entrepreneurship (Chicago: University of
Chicago Press, 1973).
Kleene, Stephen C., “Origins of Recursive Function Theory,” in Symposium on the
Foundations of Computer Science, proceedings of the Twentieth Annual Sym¬
posium on Foundations of Computer Science, held [in San Juan, Puerto Rico] by
the IEEE Computer Society’s Technical Committee on Mathematical Founda¬
tions of Computing (New York: Institute of Electrical and Electronics En¬
gineers, 1979), pp. 371-381.
Klima, Edward S., and Bellugi, Ursula, The Signs of Language (Cambridge,
Mass.: Harvard University Press, 1979).
Kling, Rob, “Social Analyses of Computing: Theoretical Perspectives in Recent
Empirical Research,” in Computing Surveys, vol. 12 (Mar. 1980), pp. 61-110.
Klir, George Jiri, An Approach to General System Theory (New York: Van Nos¬
trand Reinhold, 1969).
Klir, George Jiri, “0n the Relation Between Cybernetics and General Systems
Theory,” in Rose, John, ed., Progress of Cybernetics, proceedings of the First
International Congress of Cybernetics, held [in London] 1969 (London and New
York: Gordon and Breach Science Publishers, 1970), vol. 1, pp. 155-165.
Knuth, Donald E., The Art of Computer Programming, Vol. 1: Fundamental Al¬
gorithms (Reading, Mass.: Addison-Wesley, 1968).
Knuth, Donald E., The Art of Computer Programming (Reading, Mass.: Addison-
Wesley, 1973-1981), 3 vols.
Knuth, Donald E., “Computer Programming as an Art” [the 1974 ACM Turing
Award Lecture], in Communications of the ACM, vol. 17 (Dec. 1974a), pp. 667-
673.
Knuth, Donald E., “Computer Science and Its Relation to Mathematics,” in Amer¬
ican Mathematical Monthly, vol. 81, no. 4 (1974b), pp. 323-343.
Kochen, Manfred, “Extension of Moore-Shannon Model for Relay Circuits,” in
IBM Journal of Research and Development, vol. 3 (Apr. 1959), pp. 169-186.
Kochen, Manfred, Some Problems in Information Science (Metuchen, N.J.: Scare¬
crow Press, 1965).
Kochen, Manfred, ed., The Growth of Knowledge: Readings on Organization and
Retrieval of Information (New York: Wiley, 1967).
Kochen, Manfred, “Stability in the Growth of Knowledge, in Saracevic, Tefko,
ed., Introduction to Information Science (New York and London: Bowker,
1970), pp. 48-55.
Kochen, Manfred, “WISE: A World Information Synthesis and Encyclopaedia,” in
Journal of Documentation, vol. 28 (Dec. 1972), pp. 322—343.
Kochen, Manfred, “Views on the Foundations of Information Science, in Debons,
Anthony, ed., Information Science: Search for Identity, proceedings of the 1972
700 CUMULATIVE LIST OF REFERENCES
Kuhn, Thomas S., “Second Thoughts on Paradigms,” in Suppe, Frederick, ed.. The
Structure of Scientific Theories (Urbana: University of Illinois Press, 1974), pp
459-482.
Kuhn, Thomas S., The Essential Tension: Selected Studies in Scientific Tradition
and Change (Chicago: University of Chicago Press, 1977).
CUMULATIVE LIST OF REFERENCES 701
Luenberger, David G., Introduction to Dynamic Systems (New York: Wiley, 1979).
Lukes, Steven,“Types of Individualism,” in Dictionary of the History of Ideas, vol.
2 (New York: Scribner’s, 1973), pp. 594-604.
MacDougall, William, Body and Mind; A History and a Defense of Animism
(Boston: Beacon Press, 1961; 1st ed., London: Methuen and Co., 1911).
Machlup, Fritz, The Production and Distribution of Knowledge in the United States
(Princeton, N.J.: Princeton University Press, 1962).
Machlup, Fritz, Essays in Economic Semantics (Englewood Cliffs, N.J.: Prentice-
Hall, 1963).
Machlup, Fritz, “Theories of the Firm: Marginalist, Behavioral, Managerial,” in
American Economic Review, vol. 57 (Mar. 1967), pp. 1-33.
Machlup, Fritz, “If Matter Could Talk,” in Morgenbesser, Sidney; Suppes, Pat¬
rick; and White, Morton, eds.. Philosophy, Science, and Method (New York:
St. Martin’s Press, 1969a), pp. 286-305. Reprinted in Machlup, Fritz, Methodol¬
ogy of Economics and Other Social Sciences (New York: Academic Press,
1978), pp. 309-332.
Machlup, Fritz, “Positive and Normative Economics: An Analysis of the Ideas,” in
Heilbroner, Robert L., ed.. Economic Means and Social Ends: Essays in Polit¬
ical Economics (Englewood Cliffs, N.J.: Prentice-Hall, 1969b), pp. 99-129. Re¬
printed in Machlup, Fritz, Methodology of Economics and Other Social Sci¬
ences (New York: Academic Press, 1978), pp. 425-450.
Machlup, Fritz, Methodology of Economics and Other Social Sciences (New York:
Academic Press, 1978).
Machlup, Fritz, “Meeting Review: An Economist’s Reflections on an Institute for
Advanced Study of Information Science,” in Journal of the American Society
for Information Science, vol. 30 (Mar. 1979a), pp. 111—113.
Machlup, Fritz, “Uses, Value, and Benefits of Knowledge,” in Knowledge, vol. 1
(Sept. 1979b), pp. 62-81.
Machlup, Fritz, Knowledge: Its Creation, Distribution, and Economic Signifi¬
cance, Volume I: Knowledge and Knowledge Production (Princeton, N.J.:
Princeton University Press, 1980).
Machlup, Fritz, Knowledge: Its Creation, Distribution, and Economic Signifi¬
cance, Volume II: The Branches of Learning (Princeton, N.J.: Princeton Uni¬
versity Press, 1982).
Machlup, Fritz, Knowledge: Its Creation, Distribution, and Economic Signifi¬
cance, Volume III: The Economics of Information and Human Capital (Prince¬
ton, N.J.: Princeton University Press, 1984).
Machlup, Fritz; Leeson, Kenneth;'and Associates,Information Through the
Printed Word: The Dissemination of Scholarly, Scientific, and Intellectual
Knowledge, Vol. 1: Book Publishing (New York: Praeger, 1978a).
Machlup, Fritz; Leeson, Kenneth; and Associates, Information Through the
Printed Word: The Dissemination of Scholarly, Scientific, and Intellectual
Knowledge, Vol. 2: Journals (New York: Praeger, 1978b).
Machlup, Fritz; Leeson, Kenneth; and Associates, Information Through the
Printed Word: The Dissemination of Scholarly, Scientific, and Intellectual
Knowledge, Vol. 3: Libraries (New York: Praeger, 1978c).
704 CUMULATIVE LIST OF REFERENCES
Mill, John Stuart, “On the Definition of Political Economy; and on the Method of
Investigation Proper to It,” in London and Westminster Review, October, 1836;
reprinted in Mill, John Stuart, Essays on Some Unsettled Questions of Political
Economy (London: John W. Parker, 1844).
Mill, John Stuart, Essays on Some Unsettled Questions of Political Economy
(London: John W. Parker, 1844; 1st ed. 1836).
Miller, George A., Language and Communication (New York: McGraw-Hill,
1951).
Miller, George A., “What Is Information Measurement?” in American Psycholo¬
gist, vol. 8 (Jan. 1953), pp. 3-11.
Miller, George A., “The Magical Number Seven, Plus or Minus Two: Some Limits
on our Capacity for Processing Information,” in Psychological Review, vol. 63
(Mar. 1956), pp. 81-97.
Miller, George A., “Some Psychological Studies of Grammar,” in American Psy¬
chologist, vol. 17 (1962), pp. 748-762.
Miller, George A., and Chomsky, Noam, “Finitary Models of Language Users,” in
Luce, Robert Duncan; Bush, Robert R.; and Galanter, Eugene, eds.. Handbook
of Mathematical Psychology, vol. 2 (New York: Wiley, 1963), chap. 13, pp. 419—
492.
Miller, George A., and Frick, Frederick C., “Statistical Behavioristics and Se¬
quences of Responses,” in Psychological Review, vol. 56 (Nov. 1949), pp. 311 —
324.
Miller, George A.; Galanter, Eugene; and Pribram, Karl H., Plans and the
Structure of Behavior (New York: Holt, Rinehart, and Winston, 1960).
Miller, George A., and Johnson-Laird, Philip N., Language and Perception (Cam¬
bridge, Mass.: Belknap Press, 1976).
Miller, James G., “The Nature of Living Systems,” in Behavioral Science, vol. 21
(Sept. 1976), pp. 295-319.
Miller, James G., Living Systems (New York: McGraw-Hill, 1978).
Miller, Jonathan, The Body in Question (New York: 1st American ed., Random
House, 1978).
Miller, Laurence, “Has Artificial Intelligence Contributed to an Understanding of
the Human Mind? A Critique of Arguments For and Against,” in Cognitive
Science, vol. 2 (Apr./June 1978), pp. 111-127.
Milsum, John H., Biological Control Systems Analysis (New York: McGraw-Hill,
1966).
Minsky, Marvin L., “Steps Toward Artificial Intelligence,” in Feigenbaum, Edward
A., and Feldman, Julian, eds.. Computers and Thought (New York: McGraw-
Hill, 1963), pp. 406-450.
Minsky, Marvin L., ed., Semantic Information Processing (Cambridge, Mass.: MIT
Press, 1968).
Minsky, Marvin L., “A Framework for Representing Knowledge,” in Winston,
Patrick H., ed., The Psychology of Computer Vision (New York: McGraw-Hill,
1975), pp. 211-277.
Minsky, Marvin L., “Computer Science and the Representation of Knowledge,” in
CUMULATIVE LIST OF REFERENCES 709
Dertouzos, Michael L., and Moses, Joel, eds., The Computer Age: A Twenty-
Year View (Cambridge, Mass.: MIT Press, 1979), pp. 392-421.
Minsky, Marvin L., and Papert, Seymour, Perceptrons: An Introduction to Compu¬
tational Geometry (Cambridge, Mass.: MIT Press, 1969).
Mittelmann, H. D., and Weber, H., eds.. Bifurcation Problems and Their Numeri¬
cal Solution. [Workshop on Bifurcation Problems and Their Numerical Solution,
University of Dortmund, Jan. 15-17, 1980] (Basel, Switzerland, and Boston:
Birkhauser, 1980).
Moivre, Abraham de, The Doctrine of Chances; or, a Method of Calculating the
Probability of Events in Play (London: 1718; 3rd ed. 1756; photo, reproduction
of 3rd ed. together with the biography by Helen M. Walker, New York: Chelsea
Press, 1967).
Monod, Jacques, Chance and Necessity: An Essay on the Natural Philosophy of
Modern Biology, trans. Austryn Wainhouse [of Le hasard et la necessite: Essai
sur la philosophic naturelle de la biologie moderne, Paris: Editions du Seuil,
1970] (New York: 1st American ed., Knopf, 1971).
Montroll, Elliott W., “Theory and Observations of the Dynamics and Statistics of
Traffic on an Open Road,” in Bogdanoff, John L., and Kozin, Frank, eds.,
Proceedings of the First Symposium on Engineering Applications of Random
Function Theory and Probability (New York: Wiley, 1963), pp. 231-269.
Montroll, Elliott W., “On the Entropy Function in Sociotechnical Systems,” in
Proceedings of the National Academy of Sciences, vol. 78 (Dec. 1981), pp.
7839-7843.
Moore, Edward F., and Shannon, Claude E., “Reliable Circuits Using Less Reli¬
able Relays,” in Journal of the Franklin Institute, part 1, vol. 262 (Sept. 1956),
pp. 191-208; part 2, vol. 262 (Oct. 1956), pp. 281-297.
Morris, Charles W., “Foundations of the Theory of Signs,” in Neurath, Otto;
Carnap, Rudolf; and Morris, Charles, eds., Foundations of the Unity of Science:
Toward an International Encyclopedia of Unified Science, vol. 1 (Chicago: Uni¬
versity of Chicago Press, 1939; bound ed., 1969), pp. 77-137.
Moser, Jungen, Stable and Random Motions in Dynamical Systems (Princeton,
N.J.: Princeton University Press, 1973).
Moses, Joel, “Symbolic Integration” (Ph.D. diss., Massachusetts Institute of Tech¬
nology, 1967).
Mountcastle, Vernon B., “An Organizing Principle for Cerebral Function: The
Unit Module and the Distributed System,” in Edelman, Gerald M., and Mount¬
castle, Vernon B., The Mindful Brain (Cambridge, Mass.: MIT Press, 1978), pp.
7-50.
Nalimov, Vasilii V., “The Penetration of the Humanities into Other Fields of
Knowledge,” in Colodny, Robert G., ed., Faces of Science (Philadelphia: ISI
Press, 1981; 1st pub. in Russian in the journal Znanie-sila, no. 5 [1979]).
Naur, Peter; Randell, Brian; and Buxton, John N., eds.,Software Engineering:
Concepts and Techniques, proceedings of the NATO conferences (New York:
Petrocelli/Charter, 1976). Part I reports on the conference held [in Garmisch,
Germany] by the NATO Science Committee, Oct. 7-11, 1968; Part II on the con¬
ference held [in Rome] by the NATO Science Committee, Oct. 27-31, 1969.
710 CUMULATIVE LIST OF REFERENCES
Wayne A., ed., Proceedings of the IFIP Congress 1965, organized by the Inter¬
national Federation for Information Processing [New York] May 24-29, 1965
(Washington, D.C.: Spartan Books; London: Macmillan, 1965), vol. 1, pp. 17-
24.
Newell, Allen; Perlis, Alan J.; and Simon, Herbert A., “Computer Science,”
letter to the editor, Science, vol. 157 (22 Sept., 1967), pp. 1373-1374.
Newell, Allen, and Shaw, John C., “Programming the Logic Theory Machine,” in
Proceedings of the Western Joint Computer Conference, sponsored by the Insti¬
tute of Radio Engineers, the American Institute of Electrical Engineers, and the
Association for Computing Machinery and held [in Los Angeles] Feb. 26-28,
1957, pp. 230-240.
Newell, Allen; Shaw, John C.; and Simon, Herbert A., “Empirical Explorations
of the Logic Theory Machine: A Case Study in Heuristics,” in Proceedings of
the Western Joint Computer Conference, sponsored by the Institute of Radio
Engineers, the American Institute of Electrical Engineers, and the Association
for Computing Machinery and held [in Los Angeles] Feb. 26-28, 1957, pp. 218—
230.
Newell, Allen, and Simon, Herbert A., Human Problem Solving (Englewood
Cliffs, N.J.: Prentice-Hall, 1972).
Newell, Allen, and Simon, Herbert A., “Computer Science as Empirical Inquiry:
Symbols and Search” [the 1976 ACM Turing Award Lecture], in Communica¬
tions of the ACM, vol. 19 (Mar. 1976a), pp. 113-126.
Newell, Allen, and Simon, Herbert A., “Symbol Manipulation,” in Encyclopedia
of Computer Science (New York: Petrocelli/Charter, 1976b), pp. 1384-1389.
Newmeyer, Frederick J., Linguistic Theory in America: The First Quarter Century
of Transformational Generative Grammar (New York: Academic Press, 1980).
Ng, David S., and Stoeckenius, Jan, “Auditing: Incentives and Truthful Report¬
ing,” in Journal of Accounting Research, vol. 17 (supplement 1979), pp. 1-24.
Nicolis, George, and Prigogine, Ilya, Self-Organization in Nonequilibrium Sys¬
tems: From Dissipative Structures to Order through Fluctuations (New York:
Wiley, 1977).
Nilsson, Nils J., Problem Solving Methods in Artificial Intelligence (New York:
McGraw-Hill, 1971).
Nilsson, Nils J., “Artificial Intelligence,” in Rosenfeld, J. L., ed., Proceedings of
the IFIP Congress, vol. 4 (New York: American Elsevier, 1974), pp. 778-801.
Nilsson, Nils J., Principles of Artificial Intelligence (Palo Alto, Calif.: Tioga Press,
1980).
Nilsson, Nils J., “Artificial Intelligence: Engineering, Science or Slogan?” in Al
Magazine, vol. 3 (Jan. 1982), pp. 2-9.
Nilsson, Nils J., and Raphael, Bertram, “Preliminary Design of an Intelligent
Robot,” in Computer and Information Sciences, vol. 2 (1967), pp. 235-259.
Nirenberg, Marshall W., “The Genetic Code: II,” in Scientific American, vol. 208
(Mar. 1963), pp. 80-94.
Nisbett, Richard, and Ross, Lee, Human Inference: Strategies and Shortcomings
of Social Judgment (Englewood Cliffs, N.J.: Prentice-Hall, 1980).
Norman, Donald A., “Memory, Knowledge, and the Answering of Questions,” in
712
CUMULATIVE LIST OF REFERENCES
Solso, Robert L., ed., Contemporary Issues in Cognitive Psychology: The Loy¬
ola Symposium on Cognitive Psychology [Chicago] 1972 (Washington, D.C.:
Winston, 1973), pp. 135-165.
Norman, Donald A., “Twelve Issues for Cognitive Science,” in Cognitive Science,
vol. 4 (Jan./Mar. 1980), pp. 1-32.
Nozick, Robert, Anarchy, State, and Utopia (New York: Basic Books, 1974).
Nyquist, Harry, “Certain Factors Affecting Telegraph Speed,” in Journal of the
American Institute of Electrical Engineers, vol. 43 (Feb. 1924), pp. 124-130.
Ochs, Sidney, Elements of Neurophysiology (New York: Wiley, 1965).
Offray de la Mettrie, Julien Jan, Man a Machine (London: 2nd ed., G. Smith,
1750; ed. with notes by Gertrude Carman Bussey, La Salle, Ill.: Open Court
Publishing Company, 1961). Translated from the French.
Ratliff, Floyd, Mach Bands: Quantitative Studies on Neural Networks in the Ret¬
ina (San Francisco: Holden-Day, 1965).
Rayward, W. Boyd, Systematic Bibliography in England: 1850-1895 (Urbana: Uni¬
versity of Illinois Graduate School of Library Science, 1967a). Occasional Paper
84.
Reilly, Park M., and Blau, G. E., “The Use of Statistical Methods to Build Mathe¬
matical Models of Chemical Reacting Systems,” in Canadian Journal of Chemi¬
cal Engineering, vol. 52 (June 1974), pp. 289-299.
Reitman, Walter R., Cognition and Thought: An Information-Processing Approach
(New York: Wiley, 1965).
Report of the Royal Commission Appointed to Inquire into the Constitution and
Government of the British Museum (London: His Majesty’s Stationery Office,
1850). Minutes of Evidence, Minute 9751.
Russell, Bertrand, Human Knowledge: Its Scope and Limits (New York: Simon
and Schuster, 1948).
Russell, Bertrand, Bertrand Russell Speaks His Mind (London: Arthur Barker,
1960).
Samuelson, Paul A., Economics (New York: McGraw-Hill, 11th ed., 1980).
Saracevic, Tefko, ed., Introduction to Information Science (New York: R R Bow-
ker, 1970).
Saussure, Ferdinand de, Cours de linguistique generate (Lausanne and Paris, 1916).
English trans., Course in General Linguistics (New York: Philosophical Li¬
brary, 1959).
Schank, Roger C., and Nash-Webber, Bonnie, eds., Theoretical Issues in Natural
Language Processing, proceedings of the first conference on Theoretical Issues
in Natural Language Processing (prepared by Bolt, Beranek and Newman, Inc.,
Cambridge, Mass., 1975).
Schank, Roger C., and Wilensky, Robert, “Response to Dresher and Hornstein,”
in Cognition, vol. 5 (June 1977), pp. 133-145.
Scheibf.l, Madge E., and Scheibel, Arnold B., “Structural Substrates for Integra¬
tive Patterns in the Brain Stem Reticular Core,” in Jasper, H. H., et al., eds.,
Reticular Formation of the Brain (Boston: Little, Brown, and Co., 1958) pp
31-68.
CUMULATIVE LIST OF REFERENCES 719
Schell, Jonathan, “The Fate of Mankind,” in The New Yorker, Feb. 1, 1982, pp. 8,
15.
Schmidt, RobertF., “Somatovisceral Sensibility,” Section 3 of Schmidt, Robert F.,
ed.. Fundamentals of Sensory Physiology, trans. Marguerite A. Biederman-
Thorson (New York: Springer-Verlag, 1978; 2nd, corrected ed., 1981), pp. 81-
125.
Schrader, Alvin M., “In Search of a Name: Information Science and Its Conceptual
Antecedents,” in Library and Information Science Research, vol. 6 (spring
1984).
Schrodinger, Erwin, What is Life? The Physical Aspect of the Living Cell (Cam¬
bridge, England: The University Press; New York: Macmillan, 1945).
Schultz, Claire, K., and Garwig, Paul L., “History of the American Documenta¬
tion Institute—A Sketch,” in American Documentation, vol. 20 (1969), pp. 152—
160.
Schumpeter, JosephA., The Theory of Economic Development, trans. Redvers Opie
(Cambridge, Mass.: Harvard University Press, 1934).
Schutz, Alfred, and Luckmann, Thomas, The Structures of the Life-World, trans.
Richard M. Zaner and A. Tristram Engelhardt, Jr. (Evanston, Ill.: Northwestern
University Press, 1973).
Science, “Research News,” in Science, vol. 212 (1981), p. 313
“Research News,” in Science, vol. 214 (1981), p. 1334
“Research News,” in Science, vol. 216 (1982), p. 969
“Research News,” in Science, vol. 217 (1982), p. 1239
Searle, John R., “Minds, Brains, and Programs” [with peer commentaries], in The
Behavioral and Brain Sciences, vol. 3 (Sept. 1980), pp. 417-457.
Searle, John R., “Minds, Brains, and Programs,” in Hofstadter, Douglas R., and
Dennett, Daniel C., eds., The Mind's I: Fantasies and Reflections on Self and
Soul (New York: Basic Books, 1981), pp. 353-373.
Searle, John R., “The Myth of the Computer.” Book review of Hofstadter, Douglas
R., and Dennett, Daniel C., eds., The Mind’s I: Reflections on Self and Soul
[New York: Basic Books, 1981] in The New York Review of Books, vol. 29 (29
Apr. 1982a), pp. 3-6.
Searle, John R., “Response to Dennett” in the letters section of The New York
Review of Books, June 24, 1982b.
Sebeok, Thomas A., Contributions to the Doctrine of Signs (Bloomington: Indiana
University and Peter de Ridder Press, 1976).
Seiffert, Helmut, Information liber die Information; Verstandigung im Alltag—
Nachrichtentechnik—Wissenschaftliches Verstehen—Informationssoziologie—
Das Wissen des Gelehrten (Mtinchen: C. H. Beck, 1968).
Sellen, Betty-Carol, ed., What Else You Can Do With A Library Degree (Syra¬
cuse, New York: Gaylord, 1980).
Shannon, Claude E., “A Mathematical Theory of Communication,” in Bell System
Technical Journal, vol. 27 (July and Oct. 1948), pp. 379—423; 623-656.
Shannon, Claude E., “Communication Theory of Secrecy Systems,” in Bell Sys¬
tem Technical Journal, vol. 28 (1949), pp. 656—715.
720 CUMULATIVE LIST OF REFERENCES
Simon, Herbert A., “Cognitive Science: The Newest Science of the Artificial,” in
Cognitive Science, vol. 4 (Jan./Mar. 1980), pp. 33-46.
Simon, Herbert A., “Studying Human Intelligence by Creating Artificial Intelli¬
gence,” in American Scientist, vol. 69 (May-June, 1981), pp. 300-309.
Simon, Herbert A., and Feigenbaum, Edward A., “An Information-Processing
Theory of Some Effects of Similarity, Familiarization, and Meaningfulness in
Verbal Learning,” in Journal of Verbal Learning and Verbal Behavior, vol. 3
(Oct. 1964), pp. 385-396.
Simon, Herbert A., and Kotovsky, Kenneth, “Human Acquisition of Concepts for
Sequential Patterns,” in Psychological Review, vol. 70 (Nov. 1963), pp. 534-
546.
Singer, Edgar A., Jr., In Search of a Way of Life (New York: Columbia University
Press, 1948).
Skinner, Burrhus F., The Behavior of Organisms (New York: Appleton-Century-
Crofts, 1938).
Skinner, Burrhus F., Verbal Behavior (New York: Appleton-Century-Crofts, 1957).
Slagle, JamesR., “A Heuristic Program that Solves Symbolic Integration Problems
in Freshman Calculus,” in Feigenbaum, Edward A., and Feldman, Julian, eds.,
Computers and Thought (New York: McGraw-Hill, 1963), pp. 191-203.
Slagle, JamesR.; Dixon, J. K.; and Jones, T. L., “List Processing Languages,” in
Encyclopedia of Computer Science (New York: Petrocelli/Charter, 1976), pp.
778-785.
Slamecka, Vladimir, “The Science and Engineering of Information,” in Finerman,
Aaron, ed.. University Education in Computing Science, proceedings of a con¬
ference on Graduate Academic and Related Research Programs in Computing
Science, held [at Stony Brook] by the State University of New York, 1967 (New
York: Academic Press, 1968), pp. 81-92.
Slamecka, Vladimir, and Gehl, John, eds., “Information Sciences at Georgia Insti¬
tute of Technology: The Formative Years 1963-1978,” in Information Process¬
ing and Management, vol. 14, no. 5 (1978), pp. 320-361.
Slamecka, Vladimir, and Pearson, Charls, “Information Science,” in Encyclope¬
dia of Computer Science (New York: Petrocelli/Charter, 1976), pp. 656-657.
Slamecka, Vladimir, and Pearson, Charls, “The Portent of Signs and Symbols,” in
Weiss, Edward C., ed., The Many Faces of Information Science (Boulder,
Colo.: Westview Press, 1977), pp. 105-128.
Slater, John C., Introduction to Chemical Physics (New York: McGraw-Hill, 1st
ed. 1938).
Sleeman, Derek E., and Brown, John Seely, eds., Intelligent Tutoring Systems
(New York: Academic Press, 1982).
Slepian, David, ed., Key Papers in the Development of Information Theory (New
York: IEEE Press, 1974).
Sloan Foundation, Advisors of the Alfred P., “Cognitive Science 1978.” Report
of the State of the Art Committee (New York: Alfred P. Sloan Foundation, Oct.
1978).
Slobin, Dan I., Psycholinguistics (Glenview, Ill.: Scott, Foresman, and Co., 1971;
2nd ed., 1979).
722 CUMULATIVE LIST OF REFERENCES
ceedings of a Symposium on the Control of Eye Movements, held [in San Fran¬
cisco] Nov. 10-11, 1969 (New York: Academic Press, 1971), pp. 363-428.
After Babel: Aspects of Language and Translation (New York and
Steiner, George,
London: Oxford University Press, 1975).
Sternberg, Saul, “High-Speed Scanning in Human Memory,” in Science, vol. 153
(5 Aug., 1966), pp. 652-654.
Sternberg, Stan R., “Language and Architecture for Parallel Image Processing,” in
Grisema, E. S., and Kanal, L. N., eds., Proceedings of the Conference on
Pattern Recognition in Practice (Amsterdam: North-Holland, 1980).
Stigler, George J., “The Economics of Information,” in Journal of Political Econ¬
omy, vol. 69 (1961), pp. 213-225.
Stirling, James, Methodus differentials sive tractatus de summatione et inter¬
polation serierum infinitarum (London: 1730; another ed., London, 1764).
Strachey, Christopher, “Computer,” in Bullock, Alan, and Stallybrass, Oliver,
eds., The Harper Dictionary of Modern Thought (New York: 1st U.S. ed.,
Harper and Row, 1977), p. 122.
Strachey, Christopher, “Computing Science,” in Bullock, Alan, and Stallybrass,
Oliver, eds., The Harper Dictionary of Modern Thought (New York: 1st U.S.
ed., Harper and Row, 1977), p. 124.
Sussman, Gerald Jay, A Computer Model of Skill Acquisition (New York: American
Elsevier, 1975).
Sutherland, Ivan E., “SKETCHPAD: A Man-Machine Graphical Communication
System,” in AFIPS Conference Proceedings, Vol. 23: 1963 Spring Joint Com¬
puter Conference, organized by the American Federation of Information Pro¬
cessing Societies and held [in Detroit] May 1963 (Washington, D.C.: Spartan
Books, 1963), pp. 329-346.
Sutton, Richard S., and Barto, Andrew G., “Towards a Modern Theory of Adap¬
tive Networks: Expectation and Prediction,” in Psychological Review, vol. 88
(1981), pp. 135-170.
Swank, Raynard C., A Unifying Influence (Metuchen, N.J.: Scarecrow Press,
1981).
Swanson, Don R., “Introduction,” in Intellectual Foundations of Library Educa¬
tion, proceedings of the twenty-ninth annual Conference of the Graduate Li¬
brary School (Chicago: University of Chicago Press, 1965), p. 2.
Swanson, Don R., “Introduction to the Conference,” in The Role of Libraries in the
Growth of Knowledge, proceedings of the fortieth Conference of the Graduate
Library School (Chicago: University of Chicago Press, 1980a).
Swanson, Don R., “Libraries and the Growth of Knowledge,” in The Role of Librar¬
ies in the Growth of Knowledge, proceedings of the fortieth Conference of the
Graduate Library School (Chicago: University of Chicago Press, 19806), pp.
112-134.
Sweeney, James L., “The Economics of Depletable Resources: Market Forces and
Intertemporal Bias,” in Review of Economic Studies, vol. 64 (Feb. 1977), pp.
125-142.
Szentagothai, Janos, and Arbib, Michael A., “Conceptual Models of Neural Or-
724 CUMULATIVE LIST OF REFERENCES
Van Gigch, John P., and Hill, Richard E., Using Systems Analysis to Implement
Cost-Effectiveness and Program Budgeting in Education (Englewood Cliffs,
N.J.: Educational Technology Publications, 1971).
Van Marlen, Geert, and Dijkstra, Auke, “Information Theory Applied to Selec¬
tion of Peaks for Retrieval of Mass Spectra,” in Analytical Chemistry, vol. 48
(1976), pp. 595-598.
Van Putten, C., and Van Schuppen, J. H., “On Stochastic Dynamical Systems,” in
Proceedings of the Fourth International Symposium on the Mathematical
Theory of Networks and Systems (Delft, Netherlands: North-Holland, 1979).
von Bertalanffy, Ludwig [see “Bertalanffy, Ludwig von”].
von Neumann, John [see “Neumann, John von”].
Wang, Hao, “Toward Mechanical Mathematics,” in IBM Journal of Research and
Development, vol. 4 (Jan. 1960), pp. 2-22.
Waples, Douglas, “The Graduate Library School at Chicago,” in The Library Quar¬
terly, vol. 1 (Jan. 1931), pp. 26-36.
Waterman, Donald A. “Generalization Learning Techniques for Automating the
Learning of Heuristics,” in Artificial Intelligence, vol. 1 (spring 1970), pp. 121—
170.
Waterman, Donald A., and Hayes-Roth, Frederick, eds., Pattern-Directed Infer¬
ence Systems. Report of a Workshop on P-D I Systems [Honolulu] 1977 (New
York: Academic Press, 1978).
Waterman, Talbot H., “Systems Theory and Biology—View of a Biologist,” in
Mesarovic, Mihajlo D., ed., Systems Theory and Biology, proceedings of the
Third Systems Symposium at Case Institute of Technology (New York: Sprin-
ger-Verlag, 1968), pp. 1-37.
Watson, James D., Molecular Biology of the Gene (New York: Benjamin, 1965).
Watson, John B., “Psychology as the Behaviorist Views It,” in Psychological Re¬
view, vol. 20 (Mar. 1913), pp. 158-177.
Weaver, Warren, “Science and Complexity,” in The American Scientist, vol. 36
(Oct. 1948), pp. 536-544.
Weaver, Warren, “The Mathematics of Information,” in Automatic Control (New
York: Simon and Schuster, 1955), pp. 100-104.
Webb, Eugene J.; Campbell, Donald T.; Schwartz, Richard D.; and Sechrest,
Lee, Unobtrusive Measures: Nonreactive Research in the Social Sciences
(Chicago: Rand McNally, 1966).
Weber, David C., “The Changing Character of the Catalog in America,” in The
Library Quarterly, vol. 34 (Jan. 1964), pp. 20-33.
Webster, Charles, ed., Samuel Hartlib and the Advancement of Learning (London:
Cambridge University Press, 1970).
Wegner, Peter, “Three Computer Cultures, Computer Technology, Computer
Mathematics, and Computer Science,” in Alt, Franz L., and Rubinoff, Morris,
eds., and Freiberger, Walter, guest editor, Advances in Computers (New York:
Academic Press, 1970), vol. 10, pp. 8-78.
Wegner, Peter, “ADA Education and Technology Transfer Activities,” in ADA
Letters (Aug. 1982).
CUMULATIVE LIST OF REFERENCES 727
Wegner, Peter, ed.; Dennis, Jack; Hammer, Michael; and Teichroew, D., assoc,
eds., Research Directions in Software Technology (Cambridge, Mass.: MIT
Press, 1979).
Weinreich, Uriel, “Semantics and Semiotics,” in International Encyclopedia of the
Social Sciences (New York: Macmillan and Free Press, 1968), vol. 14, pp. 164—
169.
Weiss, Edward C., ed., The Many Faces of Information Science (Boulder, Colo.:
Westview Press, 1977).
Weizenbaum, Joseph, Computer Power and Human Reason: From Judgment to
Calculation (San Francisco: W. H. Freeman, 1976).
Wellisch, Hans, “From Information Science to Informatics,” in Journal of Librari-
anship, vol. 4 (July 1972), pp. 164-187.
Wertheimer, Max, Drei Abhandlungen zur Gestalttheorie (Neuherausgabe, Er¬
langen: Philosophische Akademie, 1925).
Western Joint Computer Conference, Extending Man’s Intellect, proceedings of
the Western Joint Computer Conference, held [in Los Angeles] by the Institute
of Radio Engineers, the American Institute of Electrical Engineers, and the
Association for Computing Machinery, May 9-11, 1961, vol. 19.
Wheeler, William Morton, “The Ant Colony as an Organism,” in Journal of
Morphology, vol. 22, no. 2 (1911), pp. 307-325.
White, Carl M., A Historical Introduction to Library Education (Metuchen, N.J.:
Scarecrow Press, 1976).
Whitehead, Alfred North, and Russell, Bertrand, Principia Mathematica (Cam¬
bridge: Cambridge University Press, 1910-1913; 2nd ed. 1925-1927; paperback
ed. 1962).
Whittemore, Bruce J., and Yovits, Marshall C., “A Generalized Concept for the
Analysis of Information,” in Debons, Anthony, ed., Information Science:
Search for Identity, proceedings of the 1972 NATO Advanced Study Institute in
Information Science, held [in Seven Springs, Champion, Penn.] Aug. 12-20,
1972 (New York: Marcel Dekker, 1974), pp. 29-45.
Widrow, Bernard, “Pattern Recognition and Adaptive Control,” in IEEE Transac¬
tions on Applications and Industry, vol. 83 (1964), pp. 269-277.
Wiener, Norbert, Cybernetics, or Control and Communication in the Animal and
the Machine (Cambridge, Mass.: MIT Press, 1948; 2nd ed., MIT Press and
Wiley, 1961).
Wiener, Norbert, Extrapolation, Interpolation, and Smoothing of Stationary Time
Series with Engineering Applications (Cambridge, Mass.: Technology Press of
the Massachusetts Institute of Technology, 1949).
Wiener, Norbert, The Human Use of Human Beings—Cybernetics and Society
(Boston: Houghton Mifflin, 1950 and 1954; 2nd ed., Boston: Avon, 1967).
Wiener, Norbert, God and Golem, Inc.: A Comment on Certain Points Where
Cybernetics Impinges on Religion (Cambridge, Mass.: MIT Press, 1964).
Wijesinghe,O., “The Effect of Varying the Rate of Presentation on the Information
Transmission of Schizophrenic and Control Group,” in British Journal of Psy¬
chiatry, vol. 130 (May 1977), pp. 509-513.
728 CUMULATIVE LIST OF REFERENCES
Wisdom, John Oulton, “The Hypothesis of Cybernetics,” in British Journal for the
Philosophy of Science, vol. 2 (May 1951), pp. 1-24. Reprinted in Pylyshyn,
Zenon W., ed., Perspectives on the Computer Revolution (Englewood Cliffs,
N.J.: Prentice-Hall, 1970).
CUMULATIVE LIST OF REFERENCES 729
Zadeh, Lofti A., and Desoer, Charles A., Linear System Theory (New York:
McGraw-Hill, 1963).
Zadeh, Lofti A., and Polak, E., System Theory (New York: McGraw-Hill, 1969).
Zeleny, Milan, “Cybernetics and General Systems—A Unitary Science?” in Ky-
bernetes, vol. 8 (1979), pp. 17-23.
Zeleny, Milan, Autopoiesis: A Theory of Living Organization (New York: Elsevier,
1981), General Systems Research Series, vol. 3.
Ziff, Paul, “The Feelings of Robots,” in Analysis, vol. 19 (Jan. 1959), pp. 64-68.
Zimmermann, Manfred, “Neurophysiology of Sensory Systems: The Sensory Sys¬
tem in the Light of Information Theory,” in Schmidt, Robert F., ed., Funda¬
mentals of Sensory Physiology, 3rd ed. of Grundriss der Sinnesphysiologie,
trans. Marguerite A. Biederman-Thorson (New York: Springer-Verlag, 1978),
pp. 68-80.
Zloof, Moshe M., “Query-by-Example: A Data Base Language,” in International
Business Machines Systems Journal, vol. 16, no. 4 (1977), pp. 324-343.
Zmud, Robert W., “Individual Differences and MIS Success: A Review of the
Empirical Literature,” in Management Science, vol. 25 (Oct. 1979), pp 966-
979.
Zunde, Pranas, and Gehl, John, “Empirical Foundations of Information Science,”
in Williams, Martha E., ed., Annual Review of Information Science and Tech¬
nology (White Plains, N.Y.: Knowledge Industries Publications [for the Ameri¬
can Society for Information Science], 1979), vol. 14, pp. 67-92.
INDEX
Abelson, Robert P., 108, 194, 216, 249-251 Aristotle (384-322 B.C.), 10, 125, 130, 140,
Abilock, Judith, 140n, 649n 418-419, 561n, 662
Ackoff, Russell L., xv, 448n, 550, 555, 558, Armer, Paul, 255, 262
56In, 576, 603 Amol’d, V.I., 454
ADA, 168-169, 175 Arrow, Kenneth J., 583n, 630n, 649
Akaike, Hirotugu, 538n ASCII, 276, 293
Allen, Garland E., 654n Ashby, W. Ross, 442, 470, 619
Allende, President Salvador, 41, 437 Asratyan, G.E., 454
Amarel, Saul, 25, 26, 134, 233 Association of American Library Schools
Amari, Shun-ichi, 461, 464 (AALS), viii
American Association for the Advancement of Association for Computing Machinery (ACM),
Science (AAAS), vi, 373 viii, 20, 23, 164
American Association for Artificial Intelligence Association of Special Libraries and
(AAAI), vii, 101 Information Bureaus (ASLIB), 348, 399
American Association of University Professors Atkinson, Anthony A., 630n
(AAUP), vii Atlas, Ronald M., 484
American Bibliographical Society (formerly Attneave, Fred, 482, 500
Bibliographical Society of Chicago), 346 Augustinus, Aurelius [St Augustine]
American Documentation Institute (ADI), (354-430 A.D.), 561n
351-353, 372, 380-382, 399 Austrian Society for Cybernetic Studies, 429n
American Economic Association (AEA), vi, vii
American Library Association (ALA), 21, Babbage, Charles, 239, 670
345-347, 359, 367, 375-376, 399 Bach-y-Rita, Paul, 425
American Psychological Association) APA), vii Backus, John, 153, 203n
American Society for Cybernetics, 39, 457 BACON (computer system), 271-272
American Society for Information Science Bacon, Francis, 125, 665
(ASIS), 19, 352-353, 359, 372, 376, Bacon, Roger, 125
381-382, 386, 399, 403 Bailey, Herbert S., Jr., xv
Ampere, Andre-Marie, 39, 409, 459 Baiman, Stanley, 629
Anderson, James A., 102 Baker, Carl, 74n
Anderson, John R., 102, 108, 210, 289 Baker, Jeffrey J. W., 654n
Anderson, Norman H., 102 Ballard, Dana, 85
Anderson, Richard C., 102 Baralt-Torrijos, Jorge, 141
Acki, Masanao, 596n Bar-Hillel, Yehoshua, 37, 48, 53n, 495,
Aphrodite, 552 539-542, 592n, 650
APL, 168 Barker, Roger C., 64
Aplenc, A., 479 Bamea, Amir, 630n
Aquinas, Saint Thomas, 668n Barr, Avron, v, 28-31,237,249,259,264,267,
Arbib, Michael A., v, 30, 40, 74n, 81, 84-86, 274, 276, 280, 287, 289-290, 671
90, 116, 248, 434,452, 459, 463-467, 471, Bartlett, Frederic C., 86
525, 527n, 529, 653 Barto, Andrew G., 461
Archimedes (287-212 B.C.), 125 Barzun, Jacques, 584
Arden, Bruce, 66-67, 156, 165 Bateson, Gregory, 39n, 413, 561n
731
732
INDEX
Batty, David, v, 365, 399-400 Bolt, Beranek, and Newman, Inc., 74n
Bauer, Walter F., 666n Boltzmann, Ludwig von, 49,433,476,478,489,
Baumol, William J., xvi, 393, 634 500, 504-505, 510
Bayes, Thomas, 478 Bongard, Mikhail Moiseevich, 267, 269-270
Beadle, George W., 654 Bookstein, Abraham, 392
Beam, Robert D., v, 607, 612, 616-617, Boole, George, 195, 285
631-632 Booth, A. Donald, 197
Bearman, Toni Carbo, v, 365, 399^400 Borden, H., 260
Beer, Stafford, 39n, 41n, 437, 561n Borelli, Giovanni Alfonso, 420
Bell Laboratories, 17, 49, 74n, 175, 223, 381, Boulding, Kenneth E., vi, xi, 8,43n, 46,49,547,
433, 547 555,574,576, 627,634,637,643n, 659
Bellman, Richard, 433, 527n, 575 Bower, Gordon H., 108, 210, 225
Bellugi, Ursula, 308 Bowler, T. Downing, 617
Benhabib, Jess, 620 Brachman, Ronald J., 216
Bennett, James S., 259 Bradford, Samuel Clement, 392, 395
Bense, Max, 470 Branscomb, Lewis M., 50n
Bensoussan, Alain, 527n Breal, Michel Jules Alfred, 143
Bentham, Jeremy, lOn, 561n, 666-667 Bresnan, Joan, 74n, 306
Berger, J., 484 Briet, Suzanne, 362
Berlekamp, Elwyn R, 501 Brigham Young University, 356
Berlinski, David, 416n, 435, 524 Brillouin, Leon, 478
Bernard, Claude, 81-82, 193, 635, 646n British Society for International Bibliography,
Bernoulli, Daniel, 508 348
Bernstein, Nikolai, 88, 464-465 Broadbent, Donald E., 100
Bertalanffy, Ludwig von, xiii, 8,39n, 42n, 41 In, Brockett, Roger W., 525, 527n
431-432, 446, 450, 527, 555-558, 561n, Broglie, Louis de, 41, 665-666
598n
Brookes, Bertram C., 548, 645, 654n
Berwick, Robert C., v, 335, 340 Brown, Christopher, 85
Beta Phi Mu, viii Brown, John Seely, 34, 235, 647n
Bever, Thomas G., v, 297, 306, 310, 328, Brown, Robert R., 611
330, 336 Brown University, viii, 74n
Bibliographical Retrieval Services (BRS), 355 Brunei University (U.K.), 429n
Bibliographical Society, The (U.K.), 346 Bruner, Jerome S., 100, 423-424
Bibliographical Society of Chicago, 346, 347 Brunovsky, P., 527n, 539, 541
Bibliography, International Institute of, Bryson, Arthur, 596n
347-348, 352n, 399 Buchanan, Bruce G., 217, 257
Bigelow, Julian, 65, 82, 192, 410, 416, 468 Buchanan, James, 636n
Binet, Alfred, 270 Buckland, Michael K., 22
Birkel, Paul, 187n
Buckley, Walter, vi, 598, 601, 612, 616,
Bitzer, Donald, 175 631-637
Black, Max, 13n Bucy, Richard, 461-462
Blaivas, Alexander, 454
Bunge, Mario, 448, 451-452, 555, 557-559,
Blau, G.E., 484 576,585-586, 625-626
Bliss, Henry Evelyn, 368 Burks, Arthur W., 63, 69
Blum, Jay, 463 Burmeister, Edwin, 596n
Blumer, Herbert, 386 Bush, Vannevar, 373, 410, 653n
Bobrow, Daniel G., 33n, 34, 73n, 206, 215 Butler, Pierce, 350
Bobrow, Robert J., 34, 647n Butterfield, Herbert, 63
Bode, Hendrik, 607 Buxton, John N., 166
Boden, Margaret A., v, 5, 29, 229, 233, 235, Byrnes, C., 527n
287-289
Bogdanov, Aleksandr A., 446, 556-558, 626 Cadwallader, M., 612
Bohm, Stephan, 635 Caesar, Caius Julius (101-44 B.C.), 662
Bohnert, Herbert, 373 Caldwell, Samuel, 410
INDEX 733
Hofstadter, Douglas R., vi, 29, 31, 130n, Jewett, Charles C., 365, 367-368, 400
263-264, 277, 283, 285, 288,291-294 Johns, Ada Winifred, 349
Holst, Erich von, 41 In
Johns Hopkins University, 420
Holstrom, Bengt R., 630n Johnson-Laird, Philip, 234
Holzner, Burkhart, 393 Jones, T.L., 30n
Homans, George C., 611
Josiah Macy Jr. Foundation, 82, 411,
Hooke, Robert, 420 441-442
Hoos, Ida R., 552n, 56In
Jouvenel, Bertrand de, 136n
Hopf, Eberhard, 454 Julesz, Bella, 464
Homstein, Norbert, 249 Juncosa, Mario L., 666n
Howard, Ronald A., 589
Hoyle, Frederick, 386 Kailath, T., 538n
Huber, Mary Taylor, xv Kak, Avinash C., 454
Hume, David, 58, 599n
Kalman, Rudolf E., 433-434, 452, 461-462,
Huyghens, Christiaan, 126
521, 525, 527n, 529, 533-539, 552-553,
557, 568-569, 576
Icarus, 179
Kamien, Ralph, 596n
Imperial College, London, 442 Kampe de Feriet, J., 539
Indiana University, vi Kaneko, Tatsuo, 484
Ingarden, R., 539
Kant, Elaine, 187n
Inhelder, Barbel, 229
Kant, Immanuel, 38, 58, 561n, 562,598,636n,
Institute of Electrical and Electronics Engineers 663
(IEEE), 430-431, 441-442
Kaplan, Abraham, 383, 403
Institute of Information Scientists, The (UK) Kaplan, Ronald, 74n
19
Karetsky, Stephen, 400
Institute for Intellectual Cooperation, The, 401 Karmiloff-Smith, Annette, 229
Institute of Management Sciences, The (TIMS), Katz, Jerrold J., 298
vi
Kelvin, Lord, 52, 146
International Association of Cybernetics, The
Kendall, D.G., 500
429
Kent, Allen, 380, 647 n
Institut National de Recherche en Informadque
Kepler, Johannes, 557, 577
et en Automatique (ENRIA), 517n
Keyser, Samuel Jay, vi, 74n, 327, 329n, 336
International Business Machines Corporation Khinchin, Aleksandr L, 47n, 443-444
(IBM), 58, 74n, 372
Khrushchev, Nikita, 444
International Economic Association, vii Kilmer, William L., 463
International Federation for Documentation Kim, Scott, 279
(FID), 352, 380 King, Donald W., xvi, 19, 22n, 381-382
International Institute for Applied Systems King, Gilbert W., 373
Analysis (IIASA), 450 Kirchoff, Gustav Robert, 533
Itami, Hiroyuki, 630n Kirzner, Israel M., 599n
Kistiakowski, George, 156
Jackson of Burnley, Willis Jackson, Baron, 442 Kleene, Stephen C., 64, 131
Jacob, Francois, 448 Klima, Edward S., 308
Jacobi, George T., 201 Kling, Robert, 155
Jakobson, Roman, 36n, 103 Klir, George Jiri, 42n, 432, 438, 451, 527n
James, William, 225, 386 Klivington, Kenneth A., xv
Jamison, Dean, 48, 651 Knuth, Donald E„ xv, 26, 135, 156, 164, 265,
Jantsch, Erich, 56In 271,501,667
Jaspers, Karl, 621 Kochen, Manfred, vi, 47n, 140n, 371, 373n,
Jaynes, Edwin T„ 478^181, 483, 506, 513 376,401-403,453,455,457,467-468,471,
Jeffrey, Richard C., xv 643n, 647n
Jekyll and Hyde, 179 Koestler, Arthur, 63
Jensen, Michael C., 629 Koffka, Kurt, 270
Jeuck, John E., 508, 510 Kohler, Wolfgang, 207, 270
INDEX 737
367,399 464
Library of Congress, 172, 358, 372-373 Marschak, Jacob, 627, 629
Licklider, Joseph C. R., xv, 205 Marshall, John C., 84
Lieberman, Philip, 308 Martin, Clyde F., 527n
Lindsay, Peter H., 270, 425 Martin, Nathaniel F.G., 500
738 INDEX
Rapoport, Anatol, 42n, 127, 450, 451n, 601, Sattinger, D.H., 454
633n Saussure, Ferdinand de, 36n, 64
Ratliff, Floyd, 463 Sayers, W.C.B., 368, 400
Rayward, W. Boyd, viii, 22,343,346,348,350, Scalapino, D., 481
365-369, 374-376, 379, 382-383, 389, Schank, Roger C., 108, 214n, 216, 249-251,
394-397, 399-400 269-271, 281-283
Reddy, D. Raj, 270 Scheele, Carl Wilhelm, 125
Reilly, Park M., 484 Scheibel, Arnold B., 463
Reitman, Walter, 289 Scheibel, Madge E., 463
Rescher, Nicholas, 557 Schell, Jonathan, 457
Resnikoff, Howard L., xv Schlumberger-Doll, Research, 74n
Rich, Robert F., xv Schmidt, Robert F., 425, 652n
Richards, Ivor A., 142 Schrader, Alvin M., 6n
Richardson, John, 400 Schrodinger, Erwin, 111
Riemann, Georg Friedrich Bernhard, 548 Schultz, Claire K., 351
Riley, John G., 594, 629 Schumpeter, Joseph A., 599n
Rinzel, John, 465 Schupbach, Rosa, xv
Riseman, Edward M., 464 Schuster, Peter, 454
Rissanen, J., 538n Schutz, Alfred, 7n, 656n
Roberts, Lawrence G., 210, 268 Schwager, Carole, xvi
Robertson, George, 173 Schwartz, Nancy, 596n
Robinson, Alan J., 212 Science Service, 351
Rockefeller University, 454 Searle, John R., 84, 233, 235, 241, 263, 273,
Rose, John, 437 277, 280, 282
Rose, Lawrence, 140n Sears, Roebuck and Company, 507-510
Rosen, Saul, 25, 27 Sebeok, Thomas A., 36n
Rosenblatt, Frank, 199, 454n, 471 Seiffert, Helmut, 38, 48
Rosenblueth, Arturo, 65, 82, 192, 410, 416, Selkurt, Ewald E., 426
438, 453, 456, 468 Sellen, Betty-Carol, 357
Rosenbrock, H.H., 527n Senbet, Lemma W., 630n
Rosenfeld, Azriel, 454 Senn, James A., 630n
Ross, John R, 302 Shakespeare, William, 471
Ross, Lee, 288 SHAKEY (the robot), 462
Rosser, John Barkley, 64 Shannon, Claude E., xiii, xv, 5, 17, 41, 47-52,
Rothe, Carl F., 426 56, 65-67, 156, 371-372, 381, 383, 385,
Rouchaleau, Y., 527n 421, 433-434, 441-444, 455, 465, 469n,
Royal Society, 366, 399, 401, 488 475-480,483-491,493-501,505-506,510,
Ruch, Floyd Leon, 424 513,540,547,570,573,587,601,613,622,
Rugh, Wilson J., 531 646n, 658-660
Rumelhart, David E., 270 Shaw, John C., 202, 203n, 248
Russell, Bertrand, 28, 64, 66, 82, 195, 227, Shaw, Ralph, 380
238, 520, 548, 571,643, 656n, 665 Shelley, Mary Wollenstonecroft, 194
Rutgers University, 74n, 270 Shelly, Gary B., 647
Ruyer, Raymond, 59In Shera, Jesse Hauk, viii, xv, 352, 354, 379,
402-403, 649
Sabloff, Jeremy A., 620 Sherrington, Charles Scott, 82
Sager, Naomi, 140n Shik, M.L., 465
Sague, Virginia M., 19 Shils, Edward Albert, 655n
Salton, Gerard, 134 Shimoni, Y., 484
Samuel, Arthur L., 219 Shirey, Donald L., 22 n
Samuelson, Paul A., 583n, 620, 621 Shortliffe, Edward H., 217, 258, 462
San Jose State University, 430 SHRDLU, 280
Saracevic, Tefko, 155 Shwartz, Steven P., 223
Samoff, David, 260 Sigma Xj, 264
INDEX 741
Simon, Herbert A., xiii, xv, 5, 23, 25-26, 29, Steiner, George, 283
32-36, 57, 63-65, 68, 73n, 100, 104, 134, Sternberg, Saul, 100, 210
152, 164, 179n, 181, 200, 202, 210, 232, Sternberg, Stan R., 454
238-240,248,252-258,260,263-264,267, Stigleitz, Kenneth, xv
269-278,285,293-294,307,428,555,558, Stigler, George J., 627
577, 629 Stirling, James, 503-504
Simonov, P., 454 Stoeckenius, Jan, 630n
Sinai, iAkov, Grigor’evich, 500 Stoics, 36n, 139-140
Singer, Edgar A., Jr., 56In Strachey, Christopher, 23n, 152
SINTO (Sheffield Interchange Organization), Sussman, Gerald Jay, 219, 268
368 Sutherland, Ivan E., 205
SKETCHPAD, 205 Sutherland, William R., 165
Skinner, Burrhus F., 297, 319, 597 Sutton, Richard S., 461
Slagle, James R., 30n, 203 Swank, Raynard C., 388
Slamecka, Vladimir, viii, 20, 27-28, 36, 134, Swanson, Don R., xv, 344, 355, 360
140n, 141, 151, 181-183, 668n Sweeney, James L., 596n
Slater, John C., 478, 505 Swets, John Arthur, 100
Sleeman, Derek, 235 Swire, Peter, xv
Slepian, David, xv, 497 Syracuse University, 21
Sloan Foundation, 33-34, 74-75, 95, 99, 102 Szentagothai, Janos, 465
Slobin, Daniel I., 312 Szilard, Leo, 426n, 433, 469n
Sloman, Aaron, 69, 233
Small, Henry, 11, 457 Takahara, Yasuhiko, 452, 525, 527n, 556,
Smets, Philippe, 455 568-569
Smith, Adam, 599-600, 605, 635-636 Tanimoto, Taffee T., 372
Smith, Brian C., 216 Tannenbaum, A., 527n, 539
Smith, George E., 223 Tanner, J.M., 100
Smith, Reid G., 251 Tarski, Alfred, 28, 238
Snow, Charles P., 3 Tate, Vernon D., 352
Soames, Scott, 74n Taube, Mortimer, 146, 222, 380
Social Sciences and Humanities Research Technical University of Berlin, 429n
Council of Canada, 445n, 555n, 625n Teichroew, D., 648 n
Society for General Systems Research, 42,548, TEIRESIAS, 259
605 TELENET, 358
Society for Philosophy and Psychology, viii Terman, Lewis Madison, 270
Socrates (469-401 B.C.), 384 Terrace, Herbert S., 308
Solomonoff, R.J., 501 Teuber, Hans-Lukas, 426
Sommerfeld, Arnold [Johannes Wilhelm], 505 Thom, Rene, 45In, 454, 527-529, 535n,
Sontag, E., 527n 542-543, 547, 599-600n
Sonuga, J.O., 484 Thompson, Nina R., 416n
Sorenson, Paul G., 131 Thompson, Richard, 74n
Special Libraries Association, 359, 381, 399 Thue, 131
Spencer Foundation, xv Tinbergen, Jan, 537
Sperber, Daniel, 319 Tonsor, Stephen J., 386n
Sperry, Roger W., 280, 653 Torda, Clara, 252
Spinoza, Benedictus de, 56In Toussaint, Godfried, 266
Sridharan, N.S., 270, 283 Toynbee, Arnold, 655n
Stakgold, Ivar, 454 Tracy, Destutt de, 96, 124, 136-137
Stalin, Josef, 417, 444 Trappl, Robert, 429, 436
Stanford Research Institute, 32, 74n, 211 Tremblay, Jean-Paul, 131
Stanford University, v, 32, 74n, 211, 258, 388 ' Tribe, Laurence H., 552n
Stanley-Jones, Douglas, 452n Tribus, Myron, viii, 48, 372n, 475-482, 485,
Stark, Lawrence A., 425 490, 492-493, 497, 500-503, 513
Stein, Gertrude, 292 Tsetlin, M.L., 465
742 INDEX
Winograd, Terry, xv, 73n, 153, 213, 244, 249, Yockey, Hubert P., 501
268, 280-281, 315, 462 Young, John F., 416-417
Winsor, Charles, 607 Young, Robert M., 91
Winsor, Justin, 367 Yovits, Marshall C., 140n, 201, 649n
Winston, Patrick H., 219, 268
Wisdom, John Oulton, 65 Zadeh, Lofti A., 455, 527n
Wittgenstein, Ludwig von, 265, 270, 592 Zeleny, Milan, 452n, 599n
Wohler, Friedrich, 193 Zermelo, Ernst Friedrich Ferdinand, 64
Wonham, W. Murray, 527n Ziff, Paul, 216
Woods, William A., 34, 213 Zimmermann, Manfred, 425, 652n
Woodworth, Robert Sessions, 422, 423 Zloof, Moshe M., 153
Wright, H. Curtis, 140, 383-385 Zmud, Robert W., 630n
Wright, Herbert F., 64 Znaniecki, Florian, 152
Wundt, Wilhelm, 194 ZOG, 173
Zunde, Pranas, 392
Xerox Corporation, 74n Zurif, Edgar, 74n
Zwass, Vladimir, viii, 151, 179-181
Yale University, viii, 74n, 269, 281, 435 Zworykin, Vladimir, 260
date due
r
--- T"
---
O-1£-la- - -
J- - ^ t.-rVK
ncn 1 \ nm
——--
R£S£RV£ 374493
374493
ISBN 0-471-88717-X