[comp.ai.neural-nets] Tech Report Available

biafore@beowulf.ucsd.edu (Louis Steven Biafore) (03/08/89)

-----------------------------------------------------------------------
The following technical report is now available.  
-----------------------------------------------------------------------


                   DYNAMIC NODE CREATION
                             IN
                  BACKPROPAGATION NETWORKS

                         Timur Ash
                        ash@ucsd.edu


                          Abstract


     Large backpropagation (BP) networks are very  difficult
to  train.  This fact complicates the process of iteratively
testing different sized networks (i.e., networks  with  dif-
ferent  numbers of hidden layer units) to find one that pro-
vides a good mapping approximation.  This paper introduces a
new  method  called Dynamic Node Creation (DNC) that attacks
both of these issues (training large  networks  and  testing
networks with different numbers of hidden layer units).  DNC
sequentially adds nodes one at a time to the hidden layer(s)
of  the  network until the desired approximation accuracy is
achieved.  Simulation results for parity,  symmetry,  binary
addition,  and  the encoder problem are presented.  The pro-
cedure was capable of finding known  minimal  topologies  in
many  cases,  and  was  always  within  three  nodes  of the
minimum. Computational expense for finding the solutions was
comparable  to  training  normal  BP  networks with the same
final topologies.  Starting out with fewer nodes than needed
to solve the problem actually seems to help find a solution.
The method yielded a solution for every problem  tried.   BP
applied  to  the same large networks with randomized initial
weights was unable, after repeated  attempts,  to  replicate
some minimum solutions found by DNC.

-----------------------------------------------------------------------
Requests for reprints should be sent to the Institute for Cognitive 
Science, C-015; University of California, San Diego; La Jolla, CA 92093.

(ICS Report 8901)
-----------------------------------------------------------------------

kadirkam@paul.rutgers.edu (Janardhanan Kadirkamanathan) (11/22/90)

Message forwarded from visakan@eng.cam.ac.uk follows:
---------------------------------------------------------

The following technical report is available:

		f-Projections: A nonlinear recursive estimation
                     algorithm for neural networks

                 V.Kadirkamanathan, M.Niranjan & F.Fallside

		     Technical Report CUED/F-INFENG/TR.53
		 Cambridge University Engineering Department
               Trumpington Street, Cambridge CB2 1PZ, England


				Abstract

By addressing the problem of sequential learning in neural networks,
we develop a new principle, f-projection, as a general method of
choosing a posterior estimate of a function, from the new information
received and the prior estimate. The principle is based on function
approximation and the posterior so obtained is optimal in the least
L-2 norm sense. The principle strikes a parallel with minimum cross
entropy, which provides a method of choosing a posterior probability
density estimate. Some fundamental properties of the principle of 
f-projection are given with formal proofs.

Based on the principle of f-projection, a recursive (sequential)
estimation method called  the method of successive f-projections, is 
proposed. Some convergence related properties for this method are 
given with formal proofs. The method is extended for parameter 
estimation and to a sequential training algorithm for neural networks.
The problem of combining two separately trained neural networks is 
also discussed.


Please send requests to:

Visakan Kadirkamanathan
Speech Laboratory
Cambridge Unviersity Engineering Department
Trumpington Street
Cambridge CB2 1PZ
England

email: visakan@eng.cam.ac.uk

-----------------------------------------------------------------------

kadirkam@paul.rutgers.edu (Janardhanan Kadirkamanathan) (11/22/90)

Message forwarded from visakan@eng.cam.ac.uk follows:
----------------------------------------------------------

The authors in the technical report
             f-Projections: A nonlinear recursive estimation
		  algorithm for neural networks 
 should have been,
		   V.Kadirkamanathan & F.Fallside
                 Technical Report CUED/F-INFENG/TR.53
				......

A draft of the paper to appear in NIPS*90 is also available.

 	    Sequential Adaptation of Radial Basis Function Neural
            Networks and its application to time-series prediction

               V.Kadirkamanathan, M.Niranjan & F.Fallside

The complete version of the paper will appear as a technical report
shortly.

-------------------------------------------------------------------

mathew@elroy.jpl.nasa.gov (Mathew Yeates) (03/13/91)

I the following technical report (JPL Publication) is available
for anonymous ftp from the neuroprose directory at
cheops.cis.ohio-state.edu. This is a short version of a previous
paper "An Architecture With Neural Network Characteristics for Least
Squares Problems" and has appeared in various forms at several
conferences.

There are two ideas that may be of interest:
1) By making the input layer of a single layer Perceptron fully
   connected, the learning scheme approximates Newtons algorithm
   instead of steepest descent.
2) By allowing local interactions between synapses the network can
   handle time varying behavior. Specifically, the network can
   implement the Kalman Filter for estimating the state of a linear
   system.

get both yeates.pseudo-kalman.ps.Z and
         yeates.pseudo-kalman-fig.ps.Z

    A Neural Network for Computing the Pseudo-Inverse of a Matrix
            and Applications to Kalman Filtering

                     Mathew C. Yeates
            California Institute of Technology
                 Jet Propulsion Laboratory

ABSTRACT

A single layer linear neural network for associative memory is
described. The matrix which best maps a set of input keys to desired 
output targets is computed recursively by the network using a parallel
implementation of Greville's algorithm. This model differs from the 
Perceptron in that the input layer is fully interconnected leading
to a parallel approximation to Newtons algorithm. This is in contrast
to the steepest descent algorithm implemented by the Perceptron.
By further extending the model to allow synapse updates to interact
locally, a biologically plausible addition, the network implements
Kalman filtering for a single output system.