[comp.ai.neural-nets] Neuron Digest V7 #9

neuron-request@HPLMS2.HPL.HP.COM ("Neuron-Digest Moderator Peter Marvit") (02/14/91)

Neuron Digest   Wednesday, 13 Feb 1991
                Volume 7 : Issue 9

Today's Topics:
  Preprints: Speech Recognition Using a Neural Network with Time Delays
                       Technical report available
              preprint - Cooperation of Learning Algorithms
                Tech Reports AND Position Availble at USC
                    CFP - Analog VLSI Neural Networks
          A Short Course in Neural Networks and Learning Theory
     Symposium on Models of Human Identification and Categorization
                    Symposium and Forum Annoucements


Send submissions, questions, address maintenance and requests for old issues to
"neuron-request@hplabs.hp.com" or "{any backbone,uunet}!hplabs!neuron-request"
Use "ftp" to get old issues from hplpm.hpl.hp.com (15.255.176.205).

------------------------------------------------------------

Subject: Preprints: Speech Recognition Using a Neural Network with Time Delays
From:    unni@neuro.cs.gmr.com (K.P.Unnikrishnan)
Date:    Wed, 06 Feb 91 11:05:57 -0500



THE FOLLOWING PREPRINTS ARE NOW AVAILABLE:


                Speaker-Independent Digit Recognition using a
                Neural Network with Time-Delayed Connections


             K.P. Unnikrishnan     J.J. Hopfield      D.W. Tank
             GM Research Labs      Caltech            AT&T Bell Labs
             Warren, MI            Pasadena, CA       Murray Hill, NJ

                                
                                ABSTRACT

The capability of a small neural network to perform speaker-independent
recognition of spoken digits in connected speech has been investigated.
The network uses time-delays to organize rapidly changing outputs of
symbol detectors over the time scale of a word.  The network is
data-driven and unclocked.  In order to achieve useful accuracy in a
speaker-independent setting, many new ideas and procedures were
developed. These include improving the feature detectors,
self-recognition of word ends, reduction in network size, and dividing
speakers into natural classes. Quantitative experiments based on Texas
Instruments digit data bases are described.

_____________________


        Connected-Digit Speaker-Dependent Speech Recognition using a 
                Neural Network with Time-Delayed Connections
        [To appear in: IEEE Trans. on Signal Proc., March 1991]


             K.P. Unnikrishnan     J.J. Hopfield      D.W. Tank
             GM Research Labs      Caltech            AT&T Bell Labs
             Warren, MI            Pasadena, CA       Murray Hill, NJ

        
                                ABSTRACT


An analog neural network that can be taught to recognize stimulus
sequences has been used to recognize the digits in connected speech.  The
circuit computes in the analog domain, using linear circuits for signal
filtering and nonlinear circuits for simple decisions, feature extraction
and noise suppression. An analog perceptron learning rule is used to
organize the subset of connections used in the circuit that are specific
to the chosen vocabulary. Computer simulations of the learning algorithm
and circuit demonstrate recognition scores >99% for a single speaker
connected-digit database.  There is no clock; the circuit is data-driven,
and there is no necessity for endpoint detection or segmentation of the
speech signal during recognition. Training in the presence of noise
provides noise immunity up to the trained level. For the speech problem
studied here, the circuit connections need only be accurate to about 3
bits digitization depth for optimum performance. The algorithm used maps
efficiently onto analog neural network hardware: single chip
microelectronic circuits based upon this algorithm can probably be built
with current technology.

_____________________

For copies of preprints, send your mailing address to:

unni@neuro.cs.gmr.com

or

K.P.Unnikrishnan, Computer Science Department, 
GM Research Labs, Warren, MI 48090-9055

------------------------------

Subject: Technical report available
From:    stefano nolfi <IP%IRMKANT.BITNET@VMA.CC.CMU.EDU>
Date:    Wed, 06 Feb 91 13:24:05 -0400


The following technical report is now available. The paper has been
submitted to ICGA-91.

Send request to stiva@irmkant.Bitnet
e-mail comments and related references are appreciated


                        AUTO-TEACHING:
        NETWORKS THAT DEVELOP THEIR OWN TEACHING INPUT

              Stefano Nolfi       Domenico Parisi
                    Institute of Psychology
                          CNR - Rome
                  E-mail: stiva@irmkant.Bitnet


                         ABSTRACT

   Back-propagation  learning   (Rumelhart,  Hinton   and
   Williams, 1986) is a useful research tool but it has a
   number of  undesiderable features such  as  having the
   experimenter  decide  from   outside  what  should  be
   learned.  We  describe  a  number  of  simulations  of
   neural  networks  that  internally  generate their own
   teaching  input. The  networks  generate the  teaching
   input  by   trasforming   the  network  input  through
   connection weights that  are  evolved using a  form of
   genetic algorithm. What results is an innate (evolved)
   capacity not to behave efficiently  in  an environment
   but  to  learn  to behave efficiently. The analysis of
   what  these  networks  evolve  to   learn  shows  some
   interesting results.


references

Rumelhart, D.E., Hinton G.E., and   Williams, R.J. (1986).
Learning internal representations by error propagation. In
D.E.  Rumelhart,  and J.L.  McClelland,  (eds.),  Parallel
Distributed Processing.  Vol.1:  Foundations.   Cambridge,
Mass.: MIT Press.

------------------------------

Subject: preprint - Cooperation of Learning Algorithms
From:    leon%FRLRI61.BITNET@CUNYVM.CUNY.EDU,
         leon%FRLRI61.BITNET@CUNYVM.CUNY.EDU
Date:    Thu, 07 Feb 91 12:02:44 +0100


The following paper has been placed in the neuroprose archives at Ohio
State University:




                   A Framework for the Cooperation
                        of Learning Algorithms


                   Leon Bottou & Patrick Gallinari
               Laboratoire de Recherche en Informatique
                        Universite de Paris XI
                      91405 Orsay Cedex - France


Abstract

We introduce a framework for training architectures composed of several
modules. This framework, which uses a statistical formulation of learning
systems, provides a single formalism for describing many classical
connectionist algorithms as well as complex systems where several
algorithms interact. It allows to design hybrid systems which combine the
advantages of connectionist algorithms as well as other learning
algorithms.



This paper will appear in the NIPS-90 proceedings.  To retrieve it by
anonymous ftp, do the following:

unix> ftp cheops.cis.ohio-state.edu          # (or ftp 128.146.8.62)
Name (cheops.cis.ohio-state.edu:): anonymous
Password (cheops.cis.ohio-state.edu:anonymous): <ret>
ftp> cd pub/neuroprose
ftp> binary
ftp> get bottou.cooperation.ps.Z
ftp> quit
unix>
unix> zcat bottou.cooperation.ps.Z | lpr -P<your postscript printer>


------------------------------

Subject: Tech Reports AND Position Availble at USC
From:    jannaron@charlie.ece.scarolina.edu (Prof. Robert Jannarone)
Date:    Mon, 11 Feb 91 08:39:16 -0500



            T E C H N I C A L   R E P O R T S  A V A I L A B L E

     Hu, Y., Ma, K., & Jannarone, R. J. (1991).  Real-time pattern
recognition, II: visual conjunctoid neural networks.  Research Report
#NNETS91-1, Electrical & Computer Engineering Department, University of
South Carolina.

     Jannarone, R. J. (1991). Automated data processing prospects: some
examples and suggestions.  Research Report #NNETS91-5, Electrical &
Computer Engineering Department, University of South Carolina.

     Mallya, S., & Jannarone, R. J. (1991).  Real-time pattern
recognition, I: neural network algorithms for normal models.  Research
Report #NNETS91-4, Electrical & Computer Engineering Department,
University of South Carolina.

     Tatman, G., & Jannarone, R. J. (1991).  Real-time pattern
recognition, III: alternative neural networks for speech recognition.
Research Report #NNETS91-3, Electrical & Computer Engineering Department,
University of South Carolina.

     Mehta, P., & Jannarone, R. J. (1991).  Real-time neural networks,
IV: conjunctoid parallel implementation.  Research Report #NNETS91-5,
Electrical & Computer Engineering Department, University of South
Carolina.

Please mail requests to:

        Robert J. Jannarone
        Electrical and Computer Eng. Dept.
        University of South Carolina
        Columbia SC, 29208
        (803) 777-7930
        (803) 777-8045, FAX
        jannaron@ece.scarolina.edu

___________________________________________________________________________
___________________________________________________________________________


          JOB OPENING ELECTRICAL & COMPUTER ENGINEERING, TENURE TRACK,
                   UNIVERSITY OF SOUTH CAROLINA (USC)

     The USC EECE Department invites applications for tenure track faculty
positions.  Particular areas of interest include qwuantum and physical
electronics, computer architecture, and computer vision.  PERSONS OF HIGH
CALIBER IN OTHER AREAS WILL BE CONSIDERED (the EECE Department houses a
highly active research laboratory in neurocomputing).  Appointment will
be at the Assistant or Associate Professor level with a competetive
salary and rank commensurate with qualifications.  Tenured appointments
at the Professor level are also possible for uniquely qualified
individuals.  The USC, as the flagship university of the state, seeks
candidates having a strong commitment to excellence in both education and
research.  Candidates for Associate Professor are expected to have
significant research records. Candidates for Assistant Professor are
expected to show strong research potential.  Positions may be filled as
early as January, 1991 but will remain open until suitable candidates are
found.  Applicants should send resumes, including names of at least three
references, to Professor Etan Bourkoff, Chair, Department of Electrical
and Computer Engineering, Swearingen Engineering Center, University of
South Carolina, Columbia, SC 29208.  The University of South Carolina is
an equal opportunity/affirmative action employer.

------------------------------

Subject: CFP - Analog VLSI Neural Networks
From:    takefuji@axon.eeap.cwru.edu (Yoshiyasu Takefuji)
Date:    Wed, 06 Feb 91 12:47:28 -0500

                      CALL FOR PAPERS
Journal Analog Integrated Circuits and Signal Processing (AICSP)
                       Special Issue 
                            on
                 Analog VLSI Neural Networks

Papers are solicited for a special issue of the Journal Analog Integrated
Circuits and Signal Processing (AICSP) covering the growing area of
artificial neural networks to be published in September 1992.  The
special issue will cover all aspects of analog VLSI neural networks.
Topics of interest include, but are not limited to, the following:

*VLSI analog/digital systems
*Tradeoffs of analog/digital systems
*Novel applications in signal processing, optimization, and others
*Wafer scale integrated systems
*Artificial neuron/synaptic models and implementations
*On-line learning hardware.

Six copies of complete manuscripts should be sent to Yoshiyasu Takefuji
by December 15, 1991.

Guest Editor: Prof. Yoshiyasu Takefuji
        Dept. of Electrical Engineering
        Case Western Reserve University
        Cleveland, OH 44106, USA
        Phone: 216-368-6430
        Fax: 216-368-2668
        Internet: takefuji@axon.eeap.cwru.edu   

Instructions for authors can be obtained from the guest editor or by
contacting Kluwer at the following address.

Karen S. Cullen
Kluwer Academic Publishers
101 Philip Drive
Norwell, MA 02061, USA
Tel. (617) 871-6300 fax (617) 871-6528
Email Karen@world.std.com


------------------------------

Subject: A Short Course in Neural Networks and Learning Theory
From:    john@cs.rhbnc.ac.uk
Date:    Thu, 07 Feb 91 12:15:43 +0000

              ____________________________________
                        A SHORT COURSE
                              IN
               NEURAL NETWORKS AND LEARNING THEORY
                    10th and 11th April, 1991
              ____________________________________
 
Dr John Shawe-Taylor,
Department of Computer Science,
Royal Holloway and Bedford New College,
University of London,
Egham,
Surrey TW20 0EX  UK
 
 
Neural networks offer the exciting prospect of training computers to
perform tasks by example rather than explicit programming. They are
finding applications across a broad spectrum of tasks including
explosives detection, credit risk, machine vision, etc. But how reliable
are such techniques?  Can we guarantee that a machine that is programmed
by example will necessarily perform adequately in novel situations?  And
are the techniques practical for large scale applications? These
questions are currently being addressed by research in the area of
Computational Learning Theory. This theory provides invaluable insights
for assessing the risks involved in relying on a limited number of
examples as well as providing a framework for estimating the efficacy of
training methods.
 
The course will cover the main results of this theory which are needed
for the practitioner. They will permit those who are developing and using
Neural Network applications to place their performance in perspective and
realistically assess how networks will scale and how accurately they are
likely to respond to new data.
 
A key feature of the course will be its hands-on practical flavour.  It
will include sessions where participants will have an opportunity to test
out ideas in practical working examples.
 
The course covers two days:
 
Day 1: Connectionism and Neural Networks
________________________________________
An overview of connectionism stressing the main strengths and
weaknesses of the approach.  Particular emphasis will be given to areas
where the techniques are finding industrial application.  At the same
time the areas where major problems remain to be solved will be
outlined and an indication of current trends in research will be
given.
 
 
Day 2: Learning Theory for Feedforward Networks
_______________________________________________
The focus will be on applying recent advances in Computational Learning
Theory to Feedforward Neural Networks. An overview of the field of
Computational Learning Theory will be given. This theory puts training
problems in perspective and suggests effective solutions.  It also
speaks to the question of generalisation and allows predictions of
performance to be made.  The practical sessions will involve applying
these insights to the training problems of Day 1.
 
 
Who should attend?
__________________
 - Those who are involved in designing Neural Network systems or will
be required to make decisions about their application and who wish to
acquire expertise enabling them to make informed judgements about
Neural Network performance.
 
 - Those who wish to benefit from recent advances in the theoretical
understanding of Neural Networks with a view to isolating useful areas
of current research.
 
 
Each day stands alone and delegates can enrol for either one or both
days.  For more details and registration information, please write to:
 
Dr Penelope Smith,
Industrial Liaison Officer,
RHBNC,
Egham, Surrey TW20 0EX
 
or email to:
 
john@cs.rhbnc.ac.uk
 

------------------------------

Subject: Symposium on Models of Human Identification and Categorization
From:    Tony Marley <INAM%MUSICB.MCGILL.CA@BITNET.CC.CMU.EDU>
Date:    Sun, 10 Feb 91 16:27:24 -0500

                               Department of Psychology
                               McGill University
                               1205 Avenue Dr Penfield
                               Montreal
                               PQ H3A 1B1
                               Canada
 
                     February 10, 1991
 
        MODELS OF HUMAN IDENTIFICATION AND CATEGORIZATION
 
 
 Symposium at the Twenty-Fourth Annual Mathematical Psychology
                             Meeting
              A. A. J. Marley, Symposium Organizer
 
             The Society for Mathematical Psychology
      Sponsored by Indiana University, Bloomington, Indiana
                       August 10-13, 1991
 
 
At each of its Annual Meetings, the Society for Mathematical Psychology
has one or more symposia on topics of current interest. I believe that
this is an opportune time to have the proposed session since much
exciting work is being done, plus Robert Nosofsky is an organizer of the
conference at Indiana, and he has recently developed many empirical and
theoretical ideas that have encouraged others to (re)enter this area.
 
Each presenter in the symposium will have 20 to 30 minutes available to
them, plus there will be time scheduled for general discussion. This
meeting is a good place to present your theoretical ideas in detail,
although simulation and empirical results are naturally also welcome.
Remember, the Cognitive Science Society is meeting at the University of
Chicago August 7- 10, i.e. just prior to this meeting; thus, by splitting
your material in an appropriate manner between the two meetings, you will
have an excellent forum within a period of a week to present your work in
detail.
 
If you are interested in participating in this symposium, please contact
me with a TITLE and ABSTRACT. I would also be interested in suggestions
of other participants with (if possible) an email address for them.
 
To give you a clearer idea of the kind of work that I consider of direct
relevance, I mention a few researchers and some recent papers. This list
is meant to be illustrative, so please don't be annoyed if I have omitted
your favourite work (including your own).
 
                           REFERENCES
 
AHA, D. W., & MCNULTY, D. (1990). Learning attribute relevance in context
in instance-based learning algorithms. In Proceedings of the Twelfth
Annual Conference of the Cognitive Science Society.  Hillsdale, NJ:
Erlbaum.
 
ASHBY, F. G. (Ed.). (in press). Probabilistic Multidimensional Models of
Perception and Cognition. Hillsdale, NJ: Erlbaum.
 
ESTES, W. K., CAMPBELL, J. A., HATSPOULOS, N., & HURWITZ, J. B.  (1989).
Base-rate effects in category learning: A comparison of parallel network
and memory storage-retrieval models. Journal of Experimental Psychology:
Learning, Memory, and Cognition, 15, 556-571.
 
GLUCK, M. A., & BOWER, G. H. (1989). Evaluating an adaptive network model
of human learning. Journal of Memory and Language, 27, 166-195.
 
HURWITZ, J. B. (1990). A hidden-pattern network model of category
learning. Ph. D. Thesis, Department of Psychology, Harvard.
 
KRUSCKE, J. K. (1990). ALCOVE: A connectionist model of category
learning. Research Report 19, Cognitive Science, Indiana University.
 
LACOUTURE, Y., & MARLEY, A. A. J. (1990). A connectionist model of choice
and reaction time in absolute identification.  Manuscript, Universite
Laval & McGill University.
 
NOSOFSKY, R. M., & GLUCK, M. A. (1989). Adaptive networks, exemplars, and
classification rule learning. Thirtieth Annual Meeting of the Psychonomic
Society, Atlanta, Georgia.
 
RATCLIFF, R. (1990). Connectionist models of recognition memory:
Constraints imposed by learning and forgetting functions.  Psychological
Review, 97, 285-308.
 
SHEPARD, R. N. (1989). A law of generalization and connectionist
learning. Plenary Session, Eleventh Annual Conference of the Cognitive
Science Society, University of Michigan, Ann Arbor.
 
 
                 Regards
 
                     Tony
 
                 A. A. J. Marley
                 Professor
                 Director of the McGill Cognitive Science Centre

------------------------------

Subject: Symposium and Forum Annoucements
From:    jannaron@charlie.ece.scarolina.edu (Prof. Robert Jannarone)
Date:    Mon, 11 Feb 91 14:05:52 -0500


           S Y M P O S I U M   A N N O U N C E M E N T

    The 23rd Southeastern Symposium on System Theory will be held at the
University of South Carolina, Columbia, on March 10 through March 12,
1991.  Contributed papers include topics in computer structures, optics,
robotics, neural networks, pattern recognition, estimation & reliability,
circuit systems, control systems, signal processing, electromagnetics,
parallel processing, communication systems, power systems, VLSI, and
computer networks.  Invited speakers include Robert Korgon from Johns
Hopkins U.  (A History of Industry/University/Government links in
Technology Transfer).  Charles Daniels from AT&T (Fiber-Optic components
for LANs and Data Communica- tions), and Richard Edwards from
Westinghouse Savannah River Laboratories (Automatic Data Analysis
Prospects for Nuclear Waste Vitrification).

_______________________________________________________________________________
_______________________________________________________________________________

        F O R U M   A N N O U N C E M E N T


    NEUROCOMPUTING and AUTOMATIC DATA PROCESSING
      PROSPECTS for INTELLIGENT MANUFACTURING:

    A Collection of Presentations from Industry,
Government, and University Scientists, with Discussion
        (in conjunction with the Twenty-Third
       Southeastern Symposium on System Theory)

Date: March 12, 1990.
Time: 2 PM until 9 PM.
Location: The University of South Carolina (USC) Swearingen Engineering
Center

Participants:

        Russ Beckmeyer, Westinghouse Savannah River Company;
        Steven Kirk, Digital Equipment Corporation;
        Cihan H. Dagli, University of Missouri-Rolla;
        Paul Huray, USC, U.S. Office of Science and Technology Policy;
        Robert J. Jannarone, USC;
        Steven Kirk, Digital Equipment Corporation;
        Omid Omidvar, University of the District of Columbia;
        William Ranson, USC, Southeastern Manufacturing and Technology Center;
        Harold Szu, U.S. Naval Research Laboratories;
        Debra Wince-Smith, U.S. Department of Commerce;
        Paul Werbos, U.S. National Science Foundation;
        David White, Mcdonell-Douglas, Corporation.

Registration fee: $85 (dinner included)

For further information, contact: Robert Jannarone
                                  Department of Electrical and
                                    Computer Engineering
                                  University of South Carolina
                                  Columbia, SC 29208
                                  (803) 777-7930
                                  (803) 777-8045 (FAX)
                                  jannaron@ece.scarolina.edu.


------------------------------

End of Neuron Digest [Volume 7 Issue 9]
***************************************