[comp.ai.neural-nets] Neuron Digest V5 #15

neuron-request@HPLABS.HP.COM ("Neuron-Digest Moderator Peter Marvit") (03/27/89)

Neuron Digest   Sunday, 26 Mar 1989
                Volume 5 : Issue 15

Today's Topics:
                             Abracts from JETAI
                Summer course in computational neurobiology
                 Efficiency (was Re: Loan applications...)
                               Help with TSP
                             Re: Help with TSP
Unsupervised Learning Algorithms, Can you send me your comments again? [I lost
                            Neural net topology
                            Re: Data Compression
                            Re: Data Compression
                            Re: Data Compression
                            Re: Data Compression
                                light relief
                          Neural Network Programs
                  Research Opportunity for Undergraduates
                   genetic algorithms vs. backpropagation
                             noise cancellation


Send submissions, questions, address maintenance and requests for old issues to
"neuron-request@hplabs.hp.com" or "{any backbone,uunet}!hplabs!neuron-request"
ARPANET users can get old issues via ftp from hplpm.hpl.hp.com (15.255.16.205).

------------------------------------------------------------

Subject: Abracts from JETAI
From:    cfields@NMSU.Edu
Date:    Fri, 03 Mar 89 15:16:53 -0700 


The following are abstracts of papers appearing in the inaugural issue of
the Journal of Experimental and Theoretical Artificial Intelligence.  JETAI
1, 1 was published 1 January, 1989.

For submission information, please contact either of the editors:

Eric Dietrich                           Chris Fields
PACSS - Department of Philosophy        Box 30001/3CRL
SUNY Binghamton                         New Mexico State University
Binghamton, NY 13901                    Las Cruces, NM 88003-0001

dietrich@bingvaxu.cc.binghamton.edu     cfields@nmsu.edu

JETAI is published by Taylor & Francis, Ltd., London, New York, Philadelphia

_________________________________________________________________________

Minds, machines and Searle

Stevan Harnad

Behavioral & Brain Sciences, 20 Nassau Street, Princeton NJ 08542, USA

Searle's celebrated Chinese Room Argument has shaken the foundations of
Artificial Intelligence.  Many refutations have been attempted, but none
seem convincing.  This paper is an attempt to sort out explicitly the
assumptions and the logical, methodological and empirical points of
disagreement.  Searle is shown to have underestimated some features of
computer modeling, but the heart of the issue turns out to be an empirical
question about the scope and limits of the purely symbolic (computational)
model of the mind.  Nonsymbolic modeling turns out to be immune to the
Chinese Room Argument.  The issues discussed include the Total Turing Test,
modularity, neural modeling, robotics, causality and the symbol-grounding
problem.

_________________________________________________________________________

Explanation-based learning: its role in problem solving

Brent J. Krawchuck and Ian H. Witten

Knowledge Sciences Laboratory, Department of Computer Science, University of
Calgary, 2500 University Drive, NW, Calgary, Alta, Canada, T2N 1N4.

`Explanation-based' learning is a semantically-driven, knowledge-intensive
paradigm for machine learning which contrasts sharply with syntactic or
`similarity-based' approaches.  This paper redevelops the foundations of EBL
from the perspective of problem-solving.  Viewed in this light, the
technique is revealed as a simple modification to an inference engine which
gives it the ability to generalize the conditions under which the solution
to a particular problem holds.  We show how to embed generalization
invisibly within the problem solver, so that it is accomplished as inference
proceeds rather than as a separate step.  The approach is also extended to
the more complex domain of planning to illustrate that it is applicable to a
variety of logic-based problem-solvers and is by no means restricted to only
simple ones.  We argue against the current trend to isolate learning from
other activity and study it separately, preferred instead to integrate it
into the very heart of problem solving.

- ----------------------------------------------------------------------------

The recognition and classification of concepts in understanding
scientific texts

Fernando Gomez and Carlos Segami

Department of Computer Science, University of Central Florida,
Orlando, FL 32816, USA.

In understanding a novel scientific text, we may distinguish the following
processes.  First, concepts are built from the logical form of the sentence
into the final knowledge structures.  This is called concept formation.
While these concepts are being formed, they are also being recognized by
checking whether they are already in long-term memory (LTM).  Then, those
concepts which are unrecognized are integrated in LTM.  In this paper,
algorithms for the recognition and integration of concepts in understanding
scientific texts are presented.  It is shown that the integration of
concepts in scientific texts is essentially a classification task, which
determines how and where to integrate them in LTM.  In some cases, the
integration of concepts results in a reclassification of some of the
concepts already stored in LTM.  All the algorithms described here have been
implemented and are part of SNOWY, a program which reads short scientific
paragraphs and answer questions.

- ---------------------------------------------------------------------------

Exploring the No-Function-In-Structure principle

Anne Keuneke and Dean Allemang

Laboratory for Artificial Intelligence Research, Department of Computer and
Information Science, The Ohio State University, 2036 Neil Avenue Mall,
Columbus, OH 43210-1277, USA.

Although much of past work in AI has focused on compiled knowledge systems,
recent research shows renewed interest and advanced efforts both in
model-based reasoning and in the integration of this deep knowledge with
compiled problem solving structures.  Device-based reasoning can only be as
good as the model used; if the needed knowledge, correct detail, or proper
theoretical background is not accessible, performance deteriorates.  Much of
the work on model-based reasoning references the `no-function-in-structure'
principle, which was introduced be de Kleer and Brown.  Although they were
all well motivated in establishing the guideline, this paper explores the
applicability and workability of the concept as a universal principle for
model representation.  This paper first describes the principle, its intent
and the concerns it addresses.  It then questions the feasibility and the
practicality of the principle as a universal guideline for model
representation.


------------------------------

Subject: Summer course in computational neurobiology
From:    Jim Bower <jbower@bek-mc.caltech.edu>
Date:    Sun, 05 Mar 89 18:09:10 -0800 

                            
 
 Course announcement:
 
               Methods in Computational Neuroscience
   
                  The Marine Biological Laboratory
                      Woods Hole, Massachusetts
 
                     August 6 - September 2,1989


 
                        General Description
 
      The Marine Biological Laboratory (MBL) in Woods Hole Massachusetts is
a world famous marine biological laboratory that has been in existence for
over 100 years.  In addition to providing research facilities for a large
number of biologists during the summer, the MBL also sponsors a number of
outstanding courses on different topics in Biology.
        This summer will be the second year in which the MBL has offered a
course in "Methods in Computational Neuroscience".  This course is designed
as a survey of the use of computer modeling techniques in studying the
information processing capabilities of the nervous system and covers models
at all levels from biologically realistic single cells and networks of cells
to biologically relevant abstract models.  The principle aim of the course
is to provide participants with the tools to simulate the functional
properties of those neural systems of interest to them as well as to
understand the general advantages and pitfalls of this experimental
approach.
 
                  The Specific Structure of the Course 
 
      The course itself includes both a lecture series and a computer
laboratory.  The lectures are given by invited faculty whose work represents
the state of the art in computational neuroscience (see list below). The
course lecture notes have been incorporated into a book published by MIT
press (" Methods in Neuronal Modeling: From Synapses to Networks" C. Koch
and I. Segev, editors. MIT Press, Cambridge, MA.,1989).
        The computer laboratory is designed to give students hands-on
experience with the simulation techniques considered in the lecture.  It
also provides students with the opportunity to actually begin simulations of
neural systems of interest to them.  The students are guided in this effort
by the visiting lecturers and course directors, but also by several students
from the Computational Neural Systems (CNS) graduate program at Caltech who
serve as Laboratory TAs.  The lab itself consists of state of the art
graphics workstations running a GEneral NEtwork SImulation System (GENESIS)
that Dr. Bower and his colleagues at Caltech have constructed over the last
several years.  Students return to their home institutions with the GENESIS
system to continue their work.
 
                           The Students
 
        The course is designed for advanced graduate students and
postdoctoral fellows in biology, computer science, electrical engineering,
physics, or psychology with an interest in computational neuroscience.
Because of the heavy computer orientation of the Lab section, a good
computer background is required (UNIX, C or PASCAL).  In addition, students
are expected to have a basic background in neurobiology. Course enrollment
is limited to 20 so as to assure the highest quality educational experience.

                          Course Directors

 James M. Bower and Christof Koch
 Computation and Neural Systems Program
 California Institute of Technology 

                            The Faculty
 
 
 Paul Adams (Stony Brook)
 Dan Alkon (NIH)
 Richard Anderson (MIT)
 John Hildebrand (Arizona)
 John Hopfield (Caltech)
 Rudolfo Llinas (NYU)
 David Rumelhart (Stanford)
 Idan Segev (Jerusalem)
 Terrence Sejnowski (Salk/UCSD)
 David Van Essen (Caltech)
 Christoph Von der Malsburg (USC)
 
 For further information and application materials contact:
 
 Admissions Coordinator
 Marine Biological Laboratory
 Woods Hole, MA 02543
 (508) 548-3705 extension 216
 
 Application Deadline May 15, 1989
 Acceptance notification in early June.


------------------------------

Subject: Efficiency (was Re: Loan applications...)
From:    bph@buengc.BU.EDU (Blair P. Houghton)
Organization: Boston Univ. Col. of Eng.
Date:    Sat, 11 Mar 89 19:31:24 +0000 

In article <37300001@m.cs.uiuc.edu> kadie@m.cs.uiuc.edu writes:
>
>In article joe@amos.ling.ucsd.edu (Fellow Sufferer) writes:
>> 
>> Hecht-Nielsen Corp of San Diego, Ca is doing just such research.
>> Their real problem was explaining why an applicant was refused [...]
>> That's not quite as easy as it sounds.
>> 
> There is another potential problem, even if an explanation is found, it may
> be illegal.
>
> For example, the ANN may be very sensitive to zipcode. This is called
> redlining; in many places it is illegal.

This is cured easily by doing what humans should (in effect) do:

Don't allow the net to process irrelevant information.  ZIP code has nil to
do with whether an applicant will repay.  There can be no _causal_
relationship between ZIP and credit rating.  (There is a large body of
evidence supporting a positive correlation, but it's got nothing to do with
the number.)  To use it as input for a NN is to make the job _more_
difficult and _less_ accurate, whether it results in a "better class" of
clientele or not.

This raises the question of efficiency metrics for Neural Networks.  In our
example, it is bad business to lend money to deadbeats, and it is worse
business to label potentially profitable debtors as deadbeats for erroneous
reasons.  There are only so many of Donald Trump out there, thank Napoleon.
The network used to make this decision would have to be tuned to optimize
the return on the lent money.

So, like, how do you tell beforehand that it's doing its job, and that it's
not _missing_ some people who were just never allowed to have a loan before?
How do you know if a neural net is being overselective?  How do you even
define the point of overselectiveness?  It's easy in dollar-based problems:
the net with the most at the end of the game wins.  What do you use for
dollars in other situations?


                                --Blair

------------------------------

Subject: Help with TSP
From:    jal@wsu-cs.uucp (Jason Leigh)
Organization: Computer Science Department, Wayne State University
Date:    Tue, 14 Mar 89 17:02:25 +0000 

I have a problem with trying to understand how to correctly implement the
Travelling Salesman Problem (TSP) using the Boltmann machine.

I have read the pertinent parts of the PDP book and compared numerous papers
from the IEEE Conference 87 but there still seems to be something boggling
me.

>From what I understand, the objective is to minimize an energy function
that consists of components pertinent to forming a minimal permutation
matrix.

This energy function E' must be recast to a form similar to that of the
Boltzmann machine E.  This means that the weights must be designed to
present hypotheses described by E'.  What I have read seems to suggest that
the rate of E' is to be used as the weights for E.  But the question is when
we use the Boltzmann algorithm, (flipping some state Si etc..)  do the
weights need to be readjusted since the Si that pertains to E should also
pertain to E' and that would cause a change in E' and hence a change in the
weight.

Am I missing something fundamentally simple?

If any one can give me some assistance on this, perhaps elaborate on how
this is correctly done/interpreted, I would appreciate it.  Any additional
references would also be helpful.


Thanks in anticipation.

Jason Leigh

------------------------------

Subject: Re: Help with TSP
From:    marcoz@MARCOZ.BOLTZ.CS.CMU.EDU (Marco Zagha)
Organization: Carnegie-Mellon University, CS/RI
Date:    Thu, 16 Mar 89 15:01:50 +0000 

See J.J. Hopfield and D.W. Tank, "Neural Computation of Decisions in
Optimization Problems," Biological Cybernetics 52(1985) p. 141-152.

>  [...] But the question is
>  when we use the Boltzmann algorithm, (flipping some state Si etc..)
>  do the weights need to be readjusted since the Si that pertains to E
>  should also pertain to E' and that would cause a change in E' and
>  hence a change in the weight.

The weights are fixed.  Only activations are adjusted.  The Hopfield & Tank
paper explains their TSP energy function.

== Marco (marcoz@cs.cmu.edu)

------------------------------

Subject: Unsupervised Learning Algorithms, Can you send me your comments again? [I lost all replies]
From:    Francisco Camargo <camargo@cs.columbia.edu>
Date:    Thu, 16 Mar 89 14:03:10 -0500 

[Due to a mis-spelled command, I erase all replies that I had receive so far.
 Would all of you who had already sent me messages kindly do it again ? I
 appologize for the inconvenience, and appreciate the help that I'm
 getting.]


> I'm currently doing a survey article on various aspects of
> Unsupervised Learning Algorithms and would like to receive
> any references related to that area. I'm already familiar
> with most of the work done by Grossberg, Kohonen, Fukushima,
> Linsker, Ziesper, etc. I'd like pointers to comparative work
> in the area, mainly if to more traditional (non-NN) classifiers.
> In any case, any pointers are greatly appreciated.
>
> I will summarize the replies and return it to anyone who requests it.
> Please, send mail to:
>
> e-mail:   camargo@cs.columbia.edu
>
> US-mail:  Francisco A. Camargo
>           511 Computer Science Department
>           Columbia University, New York, NY, 10027
>
>  
> Tnx.                    

------------------------------

Subject: Neural net topology
From:    eghbalni@spectra.COM (Hamid Eghbalnia)
Organization: Spectragraphics, Corp., San Diego, CA
Date:    Fri, 17 Mar 89 19:51:40 +0000 


        This is to all of you who requested a summary if I got info.
        on topological study of neural nets.  There are too many of
        you for me to reply individually (thanks for all the interest).

        I hate to dissapoint everyone, but amazing as it may seem, there
        was only one reply with a reference.  

        Eugene M. Norris, 'Maximal rectangular relations', Lecture notes
        in computer science, 56, pp.475-481.

        ...!nosc!spectra!eghbalni

------------------------------

Subject: Re: Data Compression
From:    fozzard@boulder.Colorado.EDU (Richard Fozzard)
Organization: University of Colorado, Boulder
Date:    Sat, 18 Mar 89 15:56:17 +0000 

In article <10199@nsc.nsc.com> andrew@nsc.nsc.com (andrew) writes:

with regard to:
3b) Terence D. Sanger, "An Optimality Principle for Unsupervised Learning"

>   (I find it mysterious that random noise training produces orientation-
>   selective receptive fields spontaneously; what's the connection between
>   eigenvectors of an input autocorrelation and straight lines?
>   Not only are these fields similar to those found in retinal cells of
>   cat and monkey, but, as one goes down the list in order of decreasing
>   eigenvalue, resemble somewhat eigenstates of wave-functions of atoms
>   from quantum mechanics - perhaps a coincidental isomorphism!).
>

Well, I dont have the solution to the mystery, but Lehky and Sejnowski
report a similar learning of line segment receptive fields under equally
unexpected circumstances - learning surface curvature from shading.  This
work used standard back-prop instead of the Hebb rule, though.

Here's the reference:

"Neural Network Model for the Cortical Representation
of Surface Curvature from Images of Shaded Surfaces"
Sidney R. Lehky and Terrence Sejnowski
Department of Biophysics
Johns Hopkins University
In: Lund, J.S. (Ed.) Sensory Processing, Oxford (1988)


PS: If you talked to a certain Dr. Mandelbrot, he would insist that your
"coincidental isomorphism" was hardly that - remember fractals?

Richard Fozzard
University of Colorado
fozzard@boulder.colorado.edu

------------------------------

Subject: Re: Data Compression
From:    aboulang@bbn.com (Albert Boulanger)
Date:    Sat, 18 Mar 89 15:56:47 +0000 


Here is another one from my collection since I am interested in this subject:

"Dimensionality-Reduction Using Connectionist Networks"
Eric Saud, MIT AI Memo 941 (January 1987)

Also the so-called "encoder" networks using backprop where the desired
output is set to be the input (dimensionality is reduced at the hidden layer
and the hidden layer activity can serve as the desired "real" output) and
Hinton's & McClelland's recirculating networks generalization of encoder
nets (see "Learning Representations by Recirculation" Heoffrey Hinton &
James McClelland, NIPS Proceedings AIP Press 1988) can reduce
dimensionality. In general the class of learning algorithms called
"unsupervised" learning can potentially reduce dimensionality. There is
however a spectrum of characteristics among the different unsupervised
learning procedures:

Do the reduced dimensions span the space?
Are the reduced dimensions orthogonal?

Terry Sanger's algorithm does both. It would be interesting to work out what
his learning rule does with sigmoid transfer functions for the neurons.

Albert Boulanger
BBN Systems & Technologies Corporation
aboulanger@bbn.com

------------------------------

Subject: Re: Data Compression
From:    kortge@Portia.Stanford.EDU (Chris Kortge)
Organization: Stanford University
Date:    Sat, 18 Mar 89 20:04:30 +0000 

In article <10199@nsc.nsc.com> andrew@nsc.nsc.com (andrew) writes:
> [...]
>3a) Terence D. Sanger, "Optimal Unsupervised Learning in a Single-Layer
>   Linear Feedforward Neural Network", MIT AI Lab., NE43-743, Cambridge, 
>   MA 02139. TDS@wheaties.ai.mit.edu
>
>The Sanger 3a) paper is highly germane; he seems to have defined a method
>whereby maximal information preservation occurs across one layer, using
>only linear elements, and a purely local update structure. The learned
>matrix of weights becomes (row-wise) the eigenvectors of the input
>autocorrelation. [...]
>Highly relevant is his comparitive data with respect to (cf. #2) self-supervised
>backprop, where numerous criteria show GHA ("Generalised Hebbian Algorithm")
>to be superior. These criteria include:
>- training time
>  [and several other things]

It wasn't clear to me from reading Sanger's thesis that the GHA is obviously
faster than self-supervised backprop.  He says that, for backprop, "training
time still seems to be an exponential function of the number of units in the
network."  It seems like this would be problem-dependent, though, and
principal components is not that tough as typical backprop problems go.
Does anyone know of an actual scaling study on this?  (E.g., where
n-dimensional random Gaussian vectors with m known principal components are
used as inputs, and n & m are varied, keeping percent variance explainable
constant, say.)

Another problem with the claim is that the fuzzy term "training time" hides
something important.  Namely, Sanger's algorithm trains output units
one-by-one during the presentation of each pattern, and to my knowledge this
sequentiality is inherent.  Thus it could be that the GHA is superior to
backprop when measured in "pattern time", but not when measured in real time
(i.e. operations of an ideal parallel device).  Here again, I don't know the
answer; I would be interested in whatever info people have on this.

Chris Kortge

------------------------------

Subject: Re: Data Compression
From:    "Erik J. Fretheim" <efrethei@BLACKBIRD.AFIT.AF.MIL>
Organization: Air Force Institute of Technology; WPAFB, OH
Date:    19 Mar 89 00:09:49 +0000 

One more to check out is a paper by Daugman (I think) in the July 88 (or was
that August) issue of ASSP.  In it the aouthor discussed using a nueral net
to find the Gabor coefficients for an image.  Using these they were able to
represent an image in less than 1 bit per pixel.  Gabor transforms are nice
in that they have a close relationship to biological image processors.
Although the information given here is somewhat sketchy, the article
shouldn't be hard to find for anyone interested.  ejf

------------------------------

Subject: light relief
From:    andrew@nsc.nsc.com (andrew)
Organization: National Semiconductor, Santa Clara
Date:    Sun, 19 Mar 89 08:47:04 +0000 

"deep variables" of physical theory. By suitably training a (proprietary)
net with constraints given by our particular version of spacetime, he finds
that the energy minima correspond to the values of many physical constants
..pi, c, h, and so on. He calls his net "god's brain". o well.

============================================================================
        DOMAIN: andrew@logic.sc.nsc.com  
        ARPA:   nsc!logic!andrew@sun.com
        USENET: ...{amdahl,decwrl,hplabs,pyramid,sun}!nsc!logic!andrew

        Andrew Palfreyman                               408-721-4788 work
        National Semiconductor  MS D3969                408-247-0145 home
        2900 Semiconductor Dr.                  
        P.O. Box 58090                                  there's many a slip
        Santa Clara, CA  95052-8090                     'twixt cup and lip

------------------------------

Subject: Neural Network Programs
From:    David Kanecki <kanecki@vacs.uwp.wisc.edu>
Date:    Tue, 21 Mar 89 17:10:55 -0600 

             Neural Network Programs for PC's
            --------------------------------------

 
  I have developed a neural network simulator/ programmer for PC's or CP/M
machines. The system has a capacity of 2700 neurons, A and B can have a 128
neuron pattern but the multiple of the A and B neuron needs to be less than
2700.  It comes in two versions:
 
      KANN21   - Binary Neuron 0 or 1
                 2 transfer functions
                 Uses DELTA learning rule
 
      KANN30   - Continous Neuron  0<x<1
                 1 transfer function
                 Uses DELTA learning rule
                 Accurate to 4 significant digits
 
  For information please write:
 
      David Kanecki, Bio. Sci./ACS
      P.O. Box 93
      Kenosha, WI 53141
 
E-Mail: kanecki@vacs.uwp.wisc.edu

------------------------------

Subject: Research Opportunity for Undergraduates
From:    masud cader <GJTAMUSC%AUVM2.BITNET@CUNYVM.CUNY.EDU>
Date:    Thu, 23 Mar 89 16:46:10 -0400 

ANNOUNCEMENT

UNDERGRADUATE COMPUTER SCIENCE/INFORMATION SYSTEMS STUDENTS

To participate in: National Science Foundation-funded "Research Experience
for Undergraduates" at the American University, Washington, D.C.

When: Mid-May til mid-July, 1989.

Why?: To be a part of a team made up of faculty and undergraduate student
researchers working in the area of Expert Systems. The team will experiment
with software development tools, design and implement software, and develop
expert systems applications.

Students may also have the opportunity to work in other areas such as
artificial neural networks and parallel processing.

How will students be compensated?: Student participants will receive a
stipend of $2,000, plus free room and board (students MUST live on A.U.'s
campus), and be awarded three undergraduate credits. In addition, there will
be opportunities to publish papers about the research carried on in this
project.

Who is eligible?: Undergraduates with at least Junior standing (i.e., going
into Junior year or already enrolled in Junior courses) who are American
citizens are qualified. Participants must be knowledgeable about Pascal -
some 'C' experience would be nice, but not necessary.

How to apply: Send for (Part I) of application, at address given
          below, or FJ4DMUSC@AUVM2.BITNET.
          and a personal Profile (Part II).

The Profile should describe: your career goals, projects you have worked on,
subjects you enjoyed and why, and whether you are interested in
teaching/research as a career. The Profile need not be longer than one page.
Send the application and Profile to:

        Dr. Larry Medsker
        Computer Science and Information Systems Department
        The American University
        4400 Massachusetts Avenue N.W.
        Washington, D.C.  20016

For additional information, contact any of the following at
                      (202) 885-1470:

Dr. Larry Medsker, Dr. Anita La Salle, Dr. Carolyn Mc CrearyThe American
Univ Research Experience for Undergraduates

Application - Part II

Profile

Guidelines: Describe: where you are currently enrolled and your major; what
you think your career goals are; the kinds of computing projects you have
worked on; subjects you most enjoyed and why; what you think your strengths
and weaknesses are; when you expect to graduate; and, whether you are
interested in teaching and/or research as a career and why. The Profile can
be brief (one page or less).

------------------------------

Subject: genetic algorithms vs. backpropagation
From:    DMONTANA%COOPER@rcca.bbn.com
Date:    Thu, 23 Mar 89 16:37:00 -0500 


In the last edition, Arnfried Ossen wrote that "Our results indicate that
genetic algorithms cannot outperform backpropagation in feedforward networks
in general."  This is a strong statement that is almost certainly based on
misinformation about genetic algorithms.  I would like to set the record
straight and point out some important concepts for people to remember when
trying to evaluate the performance of genetic algorithms in the future.

First, all genetic algorithms are not equal.  The performance of a genetic
algorithm on a particular problem is influenced by many factors including
the method of representing individuals on chromosomes and the genetic
operators used to create new individuals.  The difference between genetic
algorithms is often many orders of magnitude.  The generic genetic algorithm
uses a binary representation and bit-based mutation and crossover operators.
It is not particularly suited to any one problem.  In our paper "Training
Feedforward Networks Using Genetic Algorithms", I and Dave Davis describe a
genetic algorithm we have developed specifically for the problem of training
neural networks which has given very good preliminary results.  Unless Ossen
actually used our genetic algorithm (which is extremely unlikely), then his
results do not reflect at all on our results.

Second, when evaluating genetic algorithms with respect to backprop, it is
important to do it on a sufficiently complex problem.  To evaluate genetic
algorithms on XOR or other toy problems is like evaluating a marathoner by
clocking him at 40 meters.  An intuitive explanation of why genetic
algorithms should be better than standard backpropagation for complex
problems but not simple ones is provided by the following analogy.  Consider
a video game where the player navigates through layers of landmines to reach
a target.  Hitting a landmine puts the player back to the beginning.  The
object is to minimize the time to the target.  Consider two strategies: (1)
to go quickly but hit a landmine in each layer with a fixed probability or
(2) to go slowly but never hit a landmine.  For a small number of layers of
landmines, strategy 1 wins.  However, the expected time of strategy 1 goes
up exponentially with the number of layers while that of strategy 2 goes up
linearly.  Now substitute backprop for strategy 1, genetic algorithms for
strategy 2, local minima for landmines, and complexity for the number of
layers of landmines, and you get the picture.  (Note that when backprop hits
a local minima, it either restarts or pops out of the local minima in a way
which loses a lot of the information which got it there.)

Third, make sure that the evaluation function (i.e. training optimality
criterion) of the genetic algorithm matches the criterion on which it will
ultimately be judged.  In particular, don't use a least-squares evaluation
function if you will judge performance on number classified correctly.
Instead, use number classified correctly as your evaluation function.
Backprop requires differentiable optimality criteria and therefore any
non-differentiable criterion must be distorted to let it work.  Genetic
algorithms should not be handicapped because of this deficiency of backprop.

David Montana
dmontana@bbn.com

------------------------------

Subject: noise cancellation
From:    andrew <amdahl!nsc!andrew@APPLE.COM>
Organization: National Semiconductor, Santa Clara
Date:    25 Mar 89 04:58:30 +0000 


Has anybody experimented with neural nets vis a vis adaptive noise
cancellation?

============================================================================
        DOMAIN: andrew@logic.sc.nsc.com  
        ARPA:   nsc!logic!andrew@sun.com
        USENET: ...{amdahl,decwrl,hplabs,pyramid,sun}!nsc!logic!andrew

        Andrew Palfreyman                               408-721-4788 work
        National Semiconductor  MS D3969                408-247-0145 home
        2900 Semiconductor Dr.                  
        P.O. Box 58090                                  there's many a slip
        Santa Clara, CA  95052-8090                     'twixt cup and lip

------------------------------

End of Neurons Digest
*********************