[net.ai] Adrressing DRogers' questions

MAREK%MIT-OZ@MIT-MC.ARPA (01/21/84)

From:  Marek W. Lugowski <MAREK%MIT-OZ@MIT-MC.ARPA>

    DROGERS (c. November '84):
      I have a few questions I would like to ask, some (perhaps most)
    essentially unanswerable at this time.

Appologies in advance for rashly attempting to answer at this time.

      - Should the initially constructed subcognitive systems be
    "learning" systems, or should they be "knowledge-rich" systems? That
    is, are the subcognitive structures implanted with their knowledge
    of the domain by the programmer, or is the domain presented to the
    system in some "pure" initial state?  Is the approach to
    subcognitive systems without learning advisable, or even possible?

I would go off on a limb and claim that attempting wholesale "learning"
first (whatever that means these days) is silly.  I would think one
would first want to spike the system with hell of a lot of knowledge
(e.g., Dughof's "Slipnet" of related concepts whose links are subject to
cummulative, partial activation which eventually makes the nodes so
connected highly relevant and therefore taken into consideration by the
system).  To repeat Minsky (and probably, most of the AI folk: one can
only learn if one already almost knows it).

      - Assuming human brains are embodiments of subcognitive systems,
    then we know how they were constructed: a very specific DNA
    blueprint controlling the paths of development possible at various
    times, with large assumptions as to the state of the intellectual
    environment.  This grand process was created by trial-and-error
    through the process of evolution, that is, essentially random
    chance. How much (if any) of the subcognitive system must be created
    essentially by random processes? If essentially all, then there are
    strict limits as to how the problem should be approached.

This is an empirical question.  If my now-attempted implementation of
the Copycat Project (which uses the Slipnet described above)
[forthcoming MIT AIM #755 by Doug Hofstadter] will converge nicely, with
trivial tweaking, I'll be inclined to hold that random processes can
indeed do most of the work.  Such is my current, unfounded, belief.  On
the other hand, a failure will not debunk my position--I could always
have messed up implementationally and made bad guesses which "threw"
the system out of its potential convergence.

      - Which processes of the human brain are essentially subcognitive
    in construction, and which use other techniques? Is this balance
    optimal?  Which structures in a computational intelligence would be
    best approached subcognitively, and which by other methods?

Won't even touch the "optimal" question.  I would guess any process
involving a great deal of fan-in would need to be subcognitive in
nature.  This is argued from efficiency.  For now, and for want of
better theories, I'd approach ALL brain functions using subcognitive
models.  The alternative to this at present means von Neumannizing the
brain, an altogether quaint thing to do...

      - How are we to judge the success of a subcognitive system? The
    problems inherent in judging the "ability" of the so-called expert
    systems will be many times worse in this area. Without specific goal
    criteria, any results will be unsatisfying and potentially illusory
    to the watching world.

Performance and plausibility (in that order) ought to be our criteria.
Judging performance accurately, however, will continue to be difficult
as long as we are forced to use current computer architectures.
Still, if a subcognitive system converges at all on a LispM, there's no
reason to damn its performance.  Plausibility is easier to demonstrate;
one needs to keep in touch with the neurosciences to do that.

      - Where will thinking systems REALLY be more useful than (much
   refined) expert systems? I would guess that for many (most?)
   applications, expertise might be preferable to intelligence. Any
   suggestions about fields for which intelligent systems would have a
   real edge over (much improved) expert systems?

It's too early (or, too late?!) to draw such clean lines.  Perhaps REAL
thinking and expertise are much more intertwined than is currently
thought.  Anyway, there is nothing to be gained by pursuing that line of
questioning before WE learn how to explicitly organize knowledge better.


Over all, I defend pursuing things subcognitively for these reasons:

  -- Not expecting thinking to be a cleanly organized, top-down driven
  activity is minimizing one's expectations.  Compare thinking with such
  activities as cellular automata (e.g., The Game of Life) or The Iterated
  Pairwise Prisoner's Dilemma Game to convince yourself of the futility of
  top-down modeling where local rules and their iterated interactions are
  very successful at concisely describing the problem at hand.  No reason
  to expect the brain's top-level behavior to be any easier to explain
  away.

  -- AI has been spending a lot of itself on forcing a von Neumannian
  interpretation on the mind.  At CMU they have it down to an art, with
  Simon's "symbolic information processing" the nowadays proverbial Holy
  Grail.  With all due respect, I'd like to see more research devoted to
  modeling various alleged brain activities with high degree of
  parallelism and probabilistic interaction, systems where "symbols" are
  not givens but intricately invovled intermediates of computation.

  -- It has not been done carefully before and I want at least a thesis
  out of it.

                                -- Marek