[sci.lang] Reprints Available: Searle/Symbol-Grounding/Categorization

harnad@phoenix.Princeton.EDU (S. R. Harnad) (05/31/90)

The following three articles on Searle's Chinese Room Argument
and the Symbol Grounding Problem are available by anonymous
ftp from phoenix.princeton.edu in directory /a/ftp/pub/harnad
(see end of message for ftp instructions):

(1)                THE SYMBOL GROUNDING PROBLEM

                   [Physica D 1990, in press]

                    Stevan Harnad
                    Department of Psychology
                    Princeton University

ABSTRACT: There has been much discussion recently about the scope and
limits of purely symbolic models of the mind and about the proper role
of connectionism in cognitive modeling. This paper describes the
"symbol grounding problem" for a semantically interpretable symbol
system:  How can its semantic interpretation be made intrinsic to the
symbol system, rather than just parasitic on the meanings in our heads?
How can the meanings of the meaningless symbol tokens, manipulated
solely on the basis of their (arbitrary) shapes, be grounded in
anything but other meaningless symbols? The problem is analogous to
trying to learn Chinese from a Chinese/Chinese dictionary alone.

A candidate solution is sketched: Symbolic representations must be
grounded bottom-up in nonsymbolic representations of two kinds:
(1) iconic representations, which are analogs of the proximal sensory
projections of distal objects and events, and (2) categorical
representations, which are learned and innate feature-detectors that
pick out the invariant features of object and event categories from
their sensory projections. Elementary symbols are the names of these
object and event categories, assigned on the basis of their
(nonsymbolic) categorical representations. Higher-order (3) symbolic
representations, grounded in these elementary symbols, consist of
symbol strings describing category membership relations ("An
X is a Y that is Z").

Connectionism is one natural candidate for the mechanism that learns
the invariant features underlying categorical representations, thereby
connecting names to the proximal projections of the distal objects they
stand for. In this way connectionism can be seen as a complementary
component in a hybrid nonsymbolic/symbolic model of the mind, rather
than a rival to purely symbolic modeling. Such a hybrid model would not
have an autonomous symbolic "module," however; the symbolic functions
would emerge as an intrinsically "dedicated" symbol system as a
consequence of the bottom-up grounding of categories' names in their
sensory representations. Symbol manipulation would be governed not just
by the arbitrary shapes of the symbol tokens, but by the nonarbitrary
shapes of the icons and category invariants in which they are grounded.

---------------------------------------------------------------------

(2)            MINDS MACHINES AND SEARLE

      J Exper. Theor. A.I. 1(1) pp. 5 - 25 (1989)

                     Stevan Harnad
                Department of Psychology
                  Princeton University
                   Princeton NJ 08544

SUMMARY: Searle's celebrated Chinese Room Argument has shaken the
foundations of Artificial Intelligence. Many refutations have been
attempted, but none seem convincing. This paper is an attempt to sort
out explicitly the assumptions and the logical, methodological and
empirical points of disagreement. Searle is shown to have
underestimated some features of computer modeling, but the heart of the
issue turns out to be an empirical question about the scope and limits
of the purely symbolic (computational) model of the mind. Nonsymbolic
modeling turns out to be immune to the Chinese Room Argument. The
issues discussed include the Total Turing Test, modularity, neural
modeling, robotics, causality and the symbol-grounding problem.

Summary and Conclusions

Searle's provocative "Chinese Room Argument" attempted to show that the
goals of "Strong AI" are unrealizable. Proponents of Strong AI are
supposed to believe that (i) the mind is a computer program, (ii) the
brain is irrelevant, and (iii) the Turing Test is decisive. Searle's
argument is that since the programmed symbol-manipulating instructions
of a computer capable of passing the Turing Test for understanding
Chinese could always be performed instead by a person who could not
understand Chinese, the computer can hardly be said to understand
Chinese. Such "simulated" understanding, Searle argues, is not the same
as real understanding, which can only be accomplished by something that
"duplicates" the "causal powers" of the brain. In the present paper the
following points have been made:

(1) Simulation versus Implementation:

Searle fails to distinguish between the simulation of a mechanism,
which is only the formal testing of a theory, and the implementation of
a mechanism, which does duplicate causal powers. Searle's "simulation"
only simulates simulation rather than implementation. It can no more be
expected to understand than a simulated airplane can be expected to
fly. Nevertheless, a successful simulation must capture formally all
the relevant functional properties of a successful implementation.

(2) Theory-Testing versus Turing-Testing:

Searle's argument conflates theory-testing and Turing-Testing.
Computer simulations formally encode and test models for
human perceptuomotor and cognitive performance capacities; they
are the medium in which the empirical and theoretical work
is done. The Turing Test is an informal and open-ended test
of whether or not people can discriminate the performance
of the implemented simulation from that of a real human being.
In a sense, we are Turing-Testing one another all the time, in
our everyday solutions to the "other minds" problem.

(3) The Convergence Argument:

Searle fails to take underdetermination into account. All scientific
theories are underdetermined by their data; i.e., the data are
compatible with more than one theory. But as the data domain grows, the
degrees of freedom for alternative (equiparametric) theories shrink.
This "convergence" constraint applies to AI's "toy" linguistic and
robotic models too, as they approach the capacity to pass the Total
(asymptotic) Turing Test. Toy models are not modules.

(4) Brain Modeling versus Mind Modeling:

Searle also fails to appreciate that the brain itself can be understood
only through theoretical modeling, and that the boundary between brain
performance and body performance becomes arbitrary as one converges on
an asymptotic model of total human performance capacity.

(5) The Modularity Assumption: 

Searle implicitly adopts a strong, untested "modularity" assumption to
the effect that certain functional parts of human cognitive performance
capacity (such as language) can be be successfully modeled
independently of the rest (such as perceptuomotor or "robotic"
capacity). This assumption may be false for models approaching the
power and generality needed to pass the Turing Test.

(6) The Teletype Turing Test versus the Robot Turing Test: 

Foundational issues in cognitive science depend critically on the truth
or falsity of such modularity assumptions. For example, the "teletype"
(linguistic) version of the Turing Test could in principle (though not
necessarily in practice) be implemented by formal symbol-manipulation
alone (symbols in, symbols out), whereas the robot version necessarily
calls for full causal powers of interaction with the outside world
(seeing, doing AND linguistic competence).

(7) The Transducer/Effector Argument:

Prior "robot" replies to Searle have not been principled ones. They
have added on robotic requirements as an arbitrary extra constraint. A
principled "transducer/effector" counterargument, however, can be based
on the logical fact that transduction is necessarily nonsymbolic,
drawing on analog and analog-to-digital functions that can only be
simulated, but not implemented, symbolically.

(8) Robotics and Causality:

Searle's argument hence fails logically for the robot version of the
Turing Test, for in simulating it he would either have to USE its
transducers and effectors (in which case he would not be simulating all
of its functions) or he would have to BE its transducers and effectors,
in which case he would indeed be duplicating their causal powers (of
seeing and doing).

(9) Symbolic Functionalism versus Robotic Functionalism:

If symbol-manipulation ("symbolic functionalism") cannot in principle
accomplish the functions of the transducer and effector surfaces, then
there is no reason why every function in between has to be symbolic
either. Nonsymbolic function may be essential to implementing minds and
may be a crucial constituent of the functional substrate of mental
states ("robotic functionalism"): In order to work as hypothesized
(i.e., to be able to pass the Turing Test), the functionalist
"brain-in-a-vat" may have to be more than just an isolated symbolic
"understanding" module -- perhaps even hybrid analog/symbolic all the
way through, as the real brain is, with the symbols "grounded"
bottom-up in nonsymbolic representations.

(10) "Strong" versus "Weak" AI:

Finally, it is not at all clear that Searle's "Strong AI"/"Weak AI"
distinction captures all the possibilities, or is even representative
of the views of most cognitive scientists. Much of AI is in any case
concerned with making machines do intelligent things rather than with
modeling the mind.

Hence, most of Searle's argument turns out to rest on unanswered
questions about the modularity of language and the scope and limits of
the symbolic approach to modeling cognition. If the modularity
assumption turns out to be false, then a top-down symbol-manipulative
approach to explaining the mind may be completely misguided because its
symbols (and their interpretations) remain ungrounded -- not for
Searle's reasons (since Searle's argument shares the cognitive
modularity assumption with "Strong AI"), but because of the
transdsucer/effector argument (and its ramifications for the kind of
hybrid, bottom-up processing that may then turn out to be optimal, or
even essential, in between transducers and effectors). What is
undeniable is that a successful theory of cognition will have to be
computable (simulable), if not exclusively computational
(symbol-manipulative). Perhaps this is what Searle means (or ought to
mean) by "Weak AI."

------------------------------------------------------------------

      CATEGORY INDUCTION AND REPRESENTATION

   [From "Categorical Perception: The Groundwork of Cognition"
       Cambridge University Press 1987, S. Harnad, Ed.]

	  Stevan Harnad
	  Psychology Department
	  Princeton University
	  Princeton NJ 08544

SUMMARY: Categorization is a very basic cognitive activity. It is
involved in any task that calls for differential responding, from
operant discrimination to pattern recognition to naming and describing
objects and states-of-affairs. Explanations of categorization range
from nativist theories denying that any nontrivial categories are
acquired by learning to inductivist theories claiming that most
categories are learned.

"Categorical perception" (CP) is the name given to a suggestive
perceptual phenomenon that may serve as a useful model for
categorization in general: For certain perceptual categories,
within-category differences look much smaller than between-category
differences even when they are of the same size physically. For
example, in color perception, differences between reds and differences
between yellows look much smaller than equal-sized differences that
cross the red/yellow boundary; the same is true of the phoneme
categories /ba/ and /da/. Indeed, the effect of the category boundary
is not merely quantitative, but qualitative.

There have been two theories to explain CP effects. The "Whorf
Hypothesis" explains color boundary effects by proposing that language
somehow determines our view of reality. The "motor theory of speech
perception" explains phoneme boundary effects by attributing them to
the patterns of articulation required for pronunciation. Both theories
seem to raise more questions than they answer, for example: (i) How
general and pervasive are CP effects? Do they occur in other modalities
besides speech-sounds and color? (ii) Are CP effects inborn or can they
be generated by learning (and if so, how)? (iii) How are categories
internally represented? How does this representation generate
successful categorization and the CP boundary effect?

Some of the answers to these questions will have to come from ongoing
research, but the existing data do suggest a provisional model for
category formation and category representation. According to this
model, CP provides our basic or elementary categories. In acquiring a
category we learn to label or identify positive and negative instances
from a sample of confusable alternatives. Two kinds of internal
representation are built up in this learning by "acquaintance": (1) an
ICONIC representation that subserves our similarity judgments and (2)
an analog/digital feature-filter that picks out the invariant
information allowing us to categorize the instances correctly. This
second, CATEGORICAL representation is associated with the category
name. Category names then serve as the atomic symbols for a third
representational system, the (3) SYMBOLIC representations that underlie
language and that make it possible for us to learn by "description."
Connectionism is one possible mechainsm for learning the sensory
invariants underlying categorization and naming.

Among the implications of the model are (a) the "cognitive identity of
(current) indiscriminables": Categories and their representations can
only be provisional and approximate, relative to the alternatives
encountered to date, rather than "exact." There is also (b) no such
thing as an absolute "feature," only those features that are invariant
within a particular context of confusable alternatives. Contrary to
prevailing "prototype" views, however, (c) such provisionally invariant
features MUST underlie successful categorization, and must be
"sufficient" (at least in the "satisficing" sense) to subserve reliable
performance with all-or-none, bounded categories, as in CP. Finally,
the model brings out some basic limitations of the
"symbol-manipulative" approach to modeling cognition, showing how (d)
symbol meanings must be functionally grounded in nonsymbolic,
"shape-preserving" representations -- iconic and categorical ones.
Otherwise, all symbol interpretations are ungrounded and indeterminate.
This amounts to a principled call for a psychophysical (rather than a
neural) "bottom-up" approach to cognition.
-------------------------------------------------------------
To retrieve a file by ftp them from a Unix/Internet site, type:
ftp phoenix.princeton.edu

When you are asked for your login, type:
anonymous

For your password, type:
ident

then change directories with:
cd pub/harnad

Then type:
binary
(This is for retrieving compressed files.)

To show the available files, type:
ls

Next, retrieve the file you want with (for example):
get filename.Z

When you have the file(s) you want, type:
quit

Next uncompress the file with:
uncompress filename.Z

Now the file will be called, simply,
filename

Finally, read or print it.

I also maintain an email group for discussion of the symbol grounding
problem. If you want to be added to the list, please write
harnad@clarity.princeton.edu

Stevan Harnad
-- 
Stevan Harnad  Department of Psychology  Princeton University
harnad@clarity.princeton.edu       srh@flash.bellcore.com
harnad@elbereth.rutgers.edu    harnad@pucc.bitnet    (609)-921-7771