[comp.ai.digest] Turing machines and brains

GKMARH@IRISHMVS.BITNET (steven horst 219-289-9067) (07/26/88)

X-Delivery-Notice:  SMTP MAIL FROM does not correspond to sender.
Date: Fri, 22 Jul 88 17:39 EDT
To: ailist@ai.ai.mit.edu
From: steven horst 219-289-9067 <GKMARH%IRISHMVS.BITNET@MITVMA.MIT.EDU>
Subject: Turing machines and brains (long)


   I don't recall how the discussion of Turing machines and their
suitability for "imitating" brains started in recent editions of
the Digest, but perhaps the following considerations may prove
useful to someone.
    I think that you can go a long way towards answering the question
of whether Turing machines can "imitate" brains or minds by rephrasing
the question in ways that make the right sorts of distinctions.
There have been a number of attempts to provide more perspicuous
terminology for modeling/simulation/duplication of mental
processes, brains -- or for that matter any other sorts of objects,
systems or processes.  (I'll use "system" as a general term for things
one might want to model.)  Unfortunately, none of these attempts
at supplying a technical usage has become standard.  (Though personally
I am partial to the one Martin Ringle uses in his anthology of
articles on AI.)
    The GENERAL question, however, seems to be something like this:
WHAT FEATURES CAN COMPUTERS (or programs, or computers running programs,
it really doesn't matter from the perspective of functional
description) HAVE IN COMMON WITH OTHER OBJECTS, SYSTEMS AND PROCESSES?
Of course some of these properties are not relevant to projects of
modeling, simulation or creation of intelligent artifacts.  (The fact
that human brains and Apple Macintoshes running AI programs both exist
on the planet Earth, for example, is of little theoretical interest.)
A second class of similarities (and differences) may or may not be
relevant, depending on one's own research interests.  The fact that
you cannot build a universal Turing machine (because it would require
an infinite storage tape) is irrelevant to the mathematician, but
for anyone interested in fast simulations, it is important to have
a machine that is able to do the relevant processing quickly and
efficiently.  So it might be the case that a Turing machine could
run a simulation of some brain process that was suitably accurate,
but intolerably slow.
    But the really interesting questions (from the philosophical
standpoint) are about (1) what it is, in general, for a program to
count as a model of some system S, and (2) what kinds of features
Turing machines (or other computers) and brains can share.  And I
think that the general answer to these questions goes something like
this: what a successful model M has in common with the system S of
which it is a model is an abstract (mathematical) form.  What, after
all, does one do in modeling?  One examines a system S, and attempts
to analyze it into constituent parts and rules that govern the
interactions of those parts.  One then develops a program and data
structures in such a manner that every unit in the system modeled
has a corresponding unit in the model, and the state changes of
the system modeled are "tracked" by the state changes of the model.
The ideal of modeling is isomorphism between the model and the
system modeled.  But of course models are never that exact, because
the very process of abstraction that yields a general theory
requires one to treat some factors as "given" or "constant" which
are not negligible in the real world.
    With respect to brains and Turing machines, it might be helpful
to ask questions in the following order:
  (1) What brain phenomena are you interested in studying?
  (2) Can these phenomena be described by lawlike generalizations?
  (3) Can a Turing machine function in a manner truly isomorphic
      to the way the brain phenomenon functions?  (I take it that
      for many real world processes, such ideal modeling cannot be
      accomplished, because to but it crudely, most of reality
      is not digital.)
  (4) Can a Turing machine function in a manner close enough to being
      isomorphic with the way the brain processes function so as to
      useful for your research project?  (Whatever it may be....)
  (5) Just what is the relationship between corresponding parts of
      the model and the system modeled?  Most notably, is functional
      isomorphism the only relevant similarity beween model and
      system modeled, or do they have some stronger relationships
      as well- e.g., do they respond to the same kinds of input
      and produce the same kinds of output?

What is interesting from the brain theorist's point of view, I should
think, is the abstract description that program and brain process
share.  The computer is just a convenient way of trying to get at that
and testing it.  Of course some people think that minds or brains
share features with (digital) computers at some lower level as well,
but that question is best kept separate from the questions about
modeling and simulation.
     Sorry this has been so long.  I hope it proves relevant to
somebody's interests.

Steven Horst
   BITNET address....gkmarh@irishmvs
   SURFACE MAIL......Department of Philosophy
                     Notre Dame, IN  46556