[comp.ai.neural-nets] Computational Learning

schiebel@a.cs.wvu.wvnet.edu (Darrell Schiebel) (02/06/90)

Could someone recomend a "good" book at a moderate level which
provides a mathemetical framework for computational/"non-AI"
learning, covering the implications from statistics/cybernetics/
thermodynamics/etc.?  These and undoubtedly many other fields seem
to provide the theoretical core for this field of research. (?)

I would greatly appreciate any assistance.  (BTW) Does anyone know the
scope, depth, or quality of Nils Nilsson's "Mathemetical Foundations of
Learning Machines" (Morgan Kaufmann) ???

Many Thanks,

Darrell Schiebel
(schiebel@a.cs.wvu.wvnet.edu)
(UN027714@wvnvaxa.wvnet.edu)
(drs@baks.com)

cjoslyn@bingvaxu.cc.binghamton.edu (Cliff Joslyn) (02/08/90)

In article <689@h.cs.wvu.wvnet.edu> schiebel@a.cs.wvu.wvnet.edu (Darrell Schiebel) writes:
>I would greatly appreciate any assistance.  (BTW) Does anyone know the
>scope, depth, or quality of Nils Nilsson's "Mathemetical Foundations of
>Learning Machines" (Morgan Kaufmann) ???

The one I know is just called /Learning Machines/ by Nilson, and it's
out of print.  It covers classic perceptrons, classifiers, committee
machines, linear and quadratic, parametric training methods, layered
machines, etc.  It was OK, but I'm not into the stuff.
-- 
O------------------------------------------------------------------------->
| Cliff Joslyn, Cybernetician at Large, cjoslyn@bingvaxu.cc.binghamton.edu
| Systems Science, SUNY Binghamton, Box 1070, Binghamton NY 13901, USA
V All the world is biscuit shaped. . .

heck@Sunburn.Stanford.EDU (Stefan P. Heck) (02/08/90)

AUTHOR:   Nilsson, Nils J.
TITLE:    Learning machines; foundations of trainable pattern-classifying
            systems. By Nils J. Nilsson.
IMPRINT:  McGraw-Hill, 1965.
          137 p.

TOPICS:   Artificial intelligence
NOTES:    Item CSUU1729870-B (Books)   Language:       Year:


I hope this helps. It seems still available....

Stefan Heck
Symbolic Systems, Stanford

shankar@rnd.GBA.NYU.EDU (Shankar Bhattachary) (02/08/90)

The original posting asked about this in combination with classifier
systems, if I remember correctly.

In "Parallel Architectures and Neural Networks"
     (proceedings of the first Italian workshop, April 27-29, 1988),
    E.R. Caianiello, ed,
    World Scientific (1989)

there is an article:
   Classifier Systems and Neural Networks,
     M. Compiani, D. Montanari, R. Serra & G. Valastro.

This article also refers to : Network Learning, by Fogelman, Gallinari,
     Le Cun, and Thiria, to appear in :
         Machine Learning, volume 3,
          Kodratoff and Michalski, editors.

Does anyone know if this volume actually appeared? I have not been able to
track it down so far.

Also, could someone tell me where Michalski is ? I thought I knew, but I
was wrong.

Thanks.

--------------------------------------------------------------------------
Shankar Bhattacharyya, Information Systems, New York University
shankar@rnd.gba.nyu.edu
--------------------------------------------------------------------------

bill@wayback.unm.edu (william horne) (02/09/90)

What come to be know as Computational Learning Theory is actually an
offshoot of work by Valient.  In fact, the Computational Learning Theory
(COLT) Conferences, are centered around Valient's approach to learning.
An introductory reference is:

Valient, L.G., "A Theory of the Learnable", Comm. of the ACM, v27, n11
	pp. 1134-1142.

Nils Nilsson's book, Learning Machines, is kind of a precursor to
conventional pattern recognition, for which a better book is clearly

Duda & Hart, Pattern Classification and Scene Analysis

fulk@cs.rochester.edu (Mark Fulk) (02/09/90)

In article <1546@ariel.unm.edu> bill@wayback.unm.edu (william horne) writes:
>What come to be know as Computational Learning Theory is actually an
>offshoot of work by Valient.  In fact, the Computational Learning Theory
>(COLT) Conferences, are centered around Valient's approach to learning.
>An introductory reference is:
>
>Valient, L.G., "A Theory of the Learnable", Comm. of the ACM, v27, n11
>	pp. 1134-1142.
>

I am a person doing computational learning theory who work does NOT
derive from Valiant's.  In fact, I started in the field before he did.
I am definitely a member of the COLT crowd, having published and presented
at COLT, and being this year's conference chair.

Computational learning theory includes any research into the mathematics
of learning machines.  Most research derives it inspiration from the work
of E. Mark Gold and Raymond Solomonoff, in the late 60's.  Gold's seminal
paper is entitled "Language identification in the limit", and appeared
in Information and Control in 1967.

The field now includes the following strands:

Recursion-theoretic (mainly what I do), which foreswears (for the time
being, not permanently) considerations of efficiency.  Our hope is to
cast light on what might EVENTUALLY be a vigorous field of feasible
learning theory.  Our claim is that the recursion-theoretic setting,
by deliberately ignoring issues such as run-time, permits easier
examination of other issues which are at least as important.  It seems
to me that the most important issues involve the sorts of data used
by a learning machine, and the criteria for its success.

Concrete, (which I've done a bit of) which tries to find efficient mechanisms
for learning specific classes of functions or sets.  For example, how
efficiently can one learn a regular language from positive and negative
examples?  Gold proved that finding the smallest DFA to fit some data
is NP-complete, but I don't think that really answers the question
(even assuming NP != P).

Probabilistic concrete: Given some sort of random source of data, one
is interested in algorithms that probably succeed, at least in
supplying a good approximation.  For example, learn a convex set in
the plane from examples chosen randomly from the set according to the
uniform distribution on the plane restricted to the set.  (Easy: draw
a bunch of examples, form the convex hull).  This kind of learning is
often called "probably approximately correct" (pac) learning.

Valiant-style, or "distribution-independent" pac learning: One must
give an algorithm that, given ANY random source of data about a problem
it is supposed to work on, probably produces an approximately correct
answer, where the approximation is measured with the same distribution
used to produce the data.

There are also questions of learning things other than programs, for example
first-order formulae, and of very powerful kinds of data (f.o. formulae
in restricted languages are quite popular again).

Obviously I'm raving on about a favorite subject.  For more information,
consult the proceedings of the First and Second Workshop on Computational
Learning Theory, published by Morgan-Kaufman.  The literature appears
mainly in Information and Computation (formerly Information and Control),
with some articles in Journal of the ACM and Theoretical Computer Science,
as well as Journal of Computer and System Sciences.  The East German
journal EIK (Elektronische Informatik und Kybernetik) is a good source
for some of the work of people in Eastern Europe.  For a REAL introduction
to the subject, including a chance to talk to real originators like Valiant,
Manuel Blum, John Case, and Rusins Freivalds,...

   COME TO ROCHESTER AUGUST 6-8, 1990!!

COLT '90, the Third Workshop on Computational Learning Theory, will be
held in Stouffer's Rochester Plaza Hotel, Rochester, New York, on those
dates.  Reception the evening of Aug. 5 (Sunday).  Send me your address,
and I'll make sure you get registration materials (email fine, I'll email
you a form).  Expect the forms towards the end of May or early June.

EVEN BETTER: SUBMIT A PAPER.

I'll post the call for papers after this article.  We are very interested
in papers that carry out mathematical analyses of neural net learning.
We are even more interested in papers that expose fresh insights into
the nature of or criteria for learning.  If you think the old stuff is
foolish, and you have the right idea to replace it, and you can write
a good mathematical paper, SUBMIT IT!  We (the program committee, I'm
just a member) might even agree!!  Just, please, do not subject us to
pure complaints with no constructive suggestions, and do not send papers
of the "I tried it twice, and it worked both times (well sort of the first
time, but the second time REALLY)" sort.

Oh, and don't send anything to me.  John Case is the program committee
chair, and has things set up to receive papers at the University of
Delaware.  Send them there; for addresses, read the call for papers.

Incidentally, the Special Interest Groups in Automata and Computability
Theory (SIGACT), and in Artificial Intelligence (SIGART), of the
Association for Computing Machinery (ACM) are sponsors of COLT '90.

Mark Fulk
fulk@cs.rochester.edu
...!uunet!rochester!fulk
Computer Science Department
University of Rochester
Rochester, NY 14627

fulk@cs.rochester.edu (Mark Fulk) (02/09/90)

			CALL FOR PAPERS
			   COLT '90
	   Third Workshop on Computational Learning Theory
			 Rochester, NY
		   August 6 - August 8, 1990

The third workshop on Computational Learning Theory will be held in
Rochester, NY.  The conference will be jointly sponsored by SIGACT and
SIGART, and is expected to be similar in style to the previous such
workshops held at MIT and UC/Santa Cruz.  Registration at COLT '90 is open.

It is expected that most papers will present rigorous formal analyses
of theoretical issues in Machine Learning.  Possible topics include,
but are not limited to: resource and robustness analysis of learning
algorithms, general learnability and non-learnability results in new and
existing computational learning models, theoretical comparisons among
learning models, and papers that connect learning theory with work in
robotics, neural nets, pattern recognition and cryptography. R. Freivalds
(Latvian State University, Riga) has agreed to present an invited talk;
the program committee may consider more such.

Authors should submit an extended abstract that consists of:

	A) cover page with title, authors' names,
	  (postal and e-mail) addresses, and a 200 word summary.  

	B) body not longer than 10 pages in twelve-point font.  

Be sure to include a clear definition of the model used, an overview
of the results, and some discussion of their significance, including
comparison to other work. Proofs or proof sketches should be included
in the technical section.  Authors should send 10 copies of their
abstract to

	John Case
	COLT '90
	Department of Computer and Information Sciences
	103 Smith Hall
	University of Delaware
	Newark, DE  19716.

The deadline for receiving submissions is April 9, 1990.  This deadline
is FIRM.  Authors will be notified by May 22, final camera-ready papers
will be due June 18, and this deadline is ABSOLUTE.  The proceedings will
be published by Morgan-Kaufmann.  For further information about submissions
contact John Case (telephone: 302-451-2711, email: case@cis.udel.edu).

Chair and local arrangements: Mark A. Fulk (U. Rochester).

Program committee:

	    J. Case (Delaware, chair),
	    D. Angluin (Yale),
	    E. Baum (NEC Research, Princeton)
	    S. Ben David (Technion, Israel),
	    M. Fulk (U. Rochester),
	    D. Haussler (UC Santa Cruz),
 	    L. Pitt (U. Illinois),
	    R. Rivest (MIT),
	    C. Smith (Maryland),
	    S. Weinstein (U. Pennsylvania).

Note: papers that have appeared in journals or that are being submitted
to other conferences are not appropriate for submission to COLT with the
exception of papers submitted to the IEEE 30th Symposium on Foundations of 
Computer Science (FOCS).

A joint submission policy coordinated with FOCS permits authors to send
a paper to both conferences; in the event that both conferences accept the
paper, it will be published in the FOCS proceedings, the authors will be
invited to give a talk at both conferences, and a short (one-page) abstract
will be printed in the COLT proceedings.

As the FOCS decisions may be quite late, authors of dual submissions
will be asked to send the abstract with their final copy, so as to
allow the publisher to substitute the abstract upon receiving word of
the FOCS decision.

It is, of course, required that authors notify both committees of the
dual submission by including a note in the cover letters.

honavar@goat.cs.wisc.edu (Vasant Honavar) (02/11/90)

In article <2937@bingvaxu.cc.binghamton.edu> cjoslyn@bingvaxu.cc.binghamton.edu.cc.binghamton.edu (Cliff Joslyn) writes:
>
>The one I know is just called /Learning Machines/ by Nilson, and it's
>out of print.  It covers classic perceptrons, classifiers, committee
>machines, linear and quadratic, parametric training methods, layered
>machines, etc.  It was OK, but I'm not into the stuff.

"Learning Machines" is being reprinted by Morgan Kaufmann 
(available early 1990).