[comp.ai.neural-nets] Neuron Digest V6 #65

neuron-request@HPLMS2.HPL.HP.COM ("Neuron-Digest Moderator Peter Marvit") (11/09/90)

Neuron Digest   Thursday,  8 Nov 1990
                Volume 6 : Issue 65

Today's Topics:
                   Info Available on AI/CogSci Program
                        Re: Neuron Digest V6 #64
                 Transputer Implementations References?
                         For NIPS*90 Presenters!
               JNNS'90 program and the mailing list (long)
         Please consider this tutorial announcement for posting


Send submissions, questions, address maintenance and requests for old issues to
"neuron-request@hplabs.hp.com" or "{any backbone,uunet}!hplabs!neuron-request"
Use "ftp" to get old issues from hplpm.hpl.hp.com (15.255.176.205).

------------------------------------------------------------

Subject: Info Available on AI/CogSci Program
From:    al@unix.cis.pitt.edu (Alan M Lesgold)
Organization: Univ. of Pittsburgh, Comp & Info Sys
Date:    04 Nov 90 15:50:27 +0000

The University of Pittsburgh's Intelligent Systems Program is a graduate
program leading to a PhD for students interested in artificial
intelligence and cognitive science.  An interdisciplinary program, it has
strong ties to faculty in computer science, psychology, medical
informatics, linguistics, history and philosophy of science, and other
departments.  Major research projects of faculty range over a variety of
formal computational approaches to cognitive science, with substantial
interest in applied as well as purely basic research.  We welcome
inquiries from potential students interested in artificial intelligence
and cognitive science.

For more information, send a postal mailing address to
        isp@unix.cis.pitt.edu

  Alan Lesgold and Rich Thomason, Co-Directors
  Intelligent Systems Program
  University of Pittsburgh

------------------------------

Subject: Re: Neuron Digest V6 #64
From:    Russ Eberhart <RCE1%APLVM.BITNET@CORNELLC.cit.cornell.edu>
Date:    Tue, 06 Nov 90 17:27:45 -0500

In response to the request that prices be included with announcements of
new books:
       Neural Network PC Tools: A Practical Guide
       Eberhart & Dobbins, Eds.
       Published by Academic Press, October 1990
       Price: US $44.95
       Book can be ordered toll-free 1-800-321-5068 from US or Canada
         except Missouri, Alaska or Hawaii: 1-314-528-8110
       The Table of Contents for this book was posted on a previous
       issue of Neuron Digest.

------------------------------

Subject: Transputer Implementations References?
From:    Ade Miller <ASM%ASTRONOMY.PHYSICS.SOUTHAMPTON.AC.UK@pucc.PRINCETON.EDU>
Date:    Wed, 07 Nov 90 12:04:00 +0000


   I have just started a Phd in applying neural networks to medical imaging
and automated data rejection in astronomy. My main interests are back-
propagation networks; optimisation and learning, applications to the
above and implementations on parallel hardware (especially Transputers).
Does anyone have any useful references on the above, especially the
Transputer implementation side, (I have already looked at 'PDP' etc.).

                        Cheers,

                          Ade.

JANET address: ASM@UK.AC.SOTON.PHASTR

------------------------------

Subject: For NIPS*90 Presenters!
From:    nips90 <nips-90@CS.YALE.EDU>
Date:    Wed, 07 Nov 90 14:17:37 -0500


The NIPS*90 Conference is now less than 3 weeks away.

I would like to encourage all NIPS*90 presenters to consider bringing
videotapes of their work if available. We will most likely have a video
projection system set up in the main conference room for oral and poster
spotlight presenters. We will also have a limited number of televisions
and video casette players available in the hall for the poster sessions.

If you would like to present a videotape (VHS format preferred), please
contact me as soon as possible so that I can reserve a machine for your
time slot or for your poster session.

Thanks much,

John Moody
NIPS*90 Program Chair

nips@cs.yale.edu
(203)432-1200 VOICE
(203)432-0593 FAX


------------------------------

Subject: JNNS'90 program and the mailing list (long)
From:    kawahara@siva.ntt.jp
Date:    Sat, 03 Nov 90 00:12:30 +0200


I have finally compiled a list of titles presented at the first annual
conference of Japan Neural Network Society.  I hope this will give a
general aspect of neural network research activities in Japan.

If you have any questions, please contact neuro-admin@tut.ac.jp, which is
the administrating group of Japanese neural network researcher's mailing
list. The mailing list is still its infancy, and the traffic is still
very low. I hope this will change in the near future.

Hideki Kawahara
NTT Basic Research Laboratories
kawahara%siva.ntt.jp@RELAY.CS.NET

  ---------- cut here ----------

                            JNNS'90
The first annual conference of the Japan Neural Network Society
                Tamagawa University, Tokyo Japan
                     10-12 September, 1990

Presentation titles: (Original titles were Japanese. These titles were
translated into English by the original authors. Some of them, indicated
by '**', were translated by the editor of this list. 'O' stands for oral
presentation and 'P' stands for poster presentation.)

O1-1, A-BP "Another type of back propagation learning", Kazuhisa Niki
(Electro technical Lab.)

O1-2, Learning of Affine Invariance by Competitive Neural Network, Suichi
Kurogi (Division of Control Engineering, Kyushu Institute of Technology)

O1-3, ** An Optimal Network Size Selection for Generalization based on
Cross Validation, Yasuhiro Wada and Mitsuo Kawato (ATR)

O1-4, Generalizing Neural Networks using Mean Curvature, Shin Suzuki and
Hideki Kawahara (NTT Basic Research Labs.)

O1-5, Evolution of Artificial Animal having Perceptron by Genetic
Algorithm, Masanori Ichinose and Tsutomu Hoshino (Institute of
Engineering Mechanics, University of Tsukuba)

O2-1, Neural Network Model of Self-Organization of Walking Patterns in
Insects, Shinichi Kimura, Masafumi Yano and Hiroshi Shimizu(Faculty of
Pharmaceutical Science University of Tokyo)

O2-2, Learning Trajectory and Force Control of Human Arm Using
Feedback-Error-Learning Scheme, Masazumi Katayama and Mitsuo Kawato(ATR
Auditory and Visual Perception Research Laboratories)

O2-3, Formation of Optimal Hand Configuration to grasp an Object, Naohiro
Fukumura, Yoji Uno and Ryoji Suzuki(Department of Mathematical
Engineering and Information Physics, Faculty of Engineering, University
of Tokyo)

O2-4, Hybrid Control of Robotic Manipulator by Neural Network Model
(Variable learning of neural networks by Fuzzy set theory), Takanori
Shibata and Toshio Fukuda (Nagoya University) Masatoshi Tokita and
Toyokazu Mitsuda (Kisarazu Technical College)

O2-5, An Overview of Neuofuzzy System, Akira KAWAMURA, Nobuo WATANABE,
Yuri Owada, Ryusuke Masuoka and Kazuo Asakawa (Computer-based Systems
Laboratory, Fujutsu Laboratories Ltd.)

O3-1, ** Cognitive Effects caused by Random Movement of Wide-Field
Patterns and Response Characteristics of MST Cells of Macaque Monkey,
Masao Kaneko, Hiroshi Nakajima, Makoto Mizuno, Eiki Hida, Hide-aki Saito
and Minoru Tsukada (Faculty of Engineering, Tamagawa University)

O3-2, Extraction of Binocular Parallax with a Neural Network and 3-D
Surface Reconstruction, Mahito Fujii, Takayuki Ito, Toshio Nakagawa (NHK
Science and Technical Research Laboratories)

O3-3, A Model of 3-D Surface Depth Perception from Its Boundary Perceived
with Binocular Viewing, Masanori Idesawa (Rikin: The Institute of
Physical and Chemical Research)

O3-4, Theory of Information Propagation, Tetsuya Takahashi(The Institute
for Physical and Chemical Research Laboratory for Neural Networks)

O3-5, A model of the transformation of color selectivity in the monkey
visual system, Hidehiko Komatsu, Shinji Kaji, Shigeru Yamane
(Electrotechnical Laboratory Neuroscience Section), Yoshie Ideura
(Komatsu Limited)

O4-1, Magical number in cognitive map and quantization of cognitive map
shape, Terunori Mori (Electrotechnical Laboratory)

O4-2, Generative representation of symbolic information in a pattern
recognition model "holovision", Hiroshi Shimizu and Yoko Yamaguchi
(Faculty of Pharmaceutical Sciences, University of Tokyo)

O4-3, Long-Term Potentiation to Temporal Pattern Stimuli in Hippocampal
Slices, Takeshi Aihara, Minoru Tsukada and Makoto Mizuno (Faculty. of
Eng. Tamagawa Univ.), Hiroshi Kato and Haruyoshi Miyagawa(Dept. of
Physiol. Yamagata Univ.)

O4-4, Bidirectional Neural Network Model for the Generation and
Recognition of Temporal Patterns, Ryoko Futami and Nozomu Hoshimiya
(Department of Electrical Communications, Faculty of Engineering, Tohoku
University)

O4-5, Theta rhythm in hippocampus: phase control of information
circulation, Yoko Yamaguchi and Hiroshi Shimizu (Faculty of
Pharmaceutical Science University of Tokyo)

O5-1, Stability and/or instability of limit cycle memories embedded in an
asynchronous neural network model, Toshinao Ishii and Wolfgang Banzhaf
(Central Research Laboratory Mitsubishi Electric Corporation), Shigetoshi
Nara (Department of Electric and Electronic Engineering Faculty of
Engineer Okayama Univ.)

O5-2, Geometric analysis of the dynamics of associative memory networks,
Kenji Doya (Faculty of Engineering, University of Tokyo)

O5-3, On the Integration of Mapping and Relaxation, Kazuyoshi Tsutsumi
(Department of Mechanical and System Engineering Faculty of Science and
Technology Ryukoku Univ.)

P1-1, Neural network model on gustatory neurons in rat, Masaharu Adachi,
Eiko Ohshima, Kazuyuki Aihara and Makoto Kotani (Faculty of Engineering,
Tokyo Denki Univ.), Takanori Nagai (Faculty of Medicine, Teikyo Univ.),
Takashi Yamamoto (Faculty of Dentistry, Osaka Univ.)

P1-2, Learning Algorithm based on Temporal Pattern Discrimination in
Hippocampus, Minoru Tsukada and Takeshi Aihara (Tamagawa University), K.
Kato (Yamagata University)

P1-3, A study on Learning by Synapse patch group, Shuji Akiyama, Yukifumi
Shigematsu and Gen Matsumoto (Electrotechnical Laboratory Molecular and
Cellular Neuroscience Section)

P1-4, A Model of the Mechanisms of Long-Term Depression in the
Cerebellum, Tatso Kitajima and Kenichi Hara (Faculty of Engineering,
Yamagata Univ.)

P1-5, ** Self-Organization in Neural Networks with Lateral Inhibition, Y.
Tamori, S. Inawashiro and Y. Musya (Faculty of Engineering, Tohoku
University)

P1-6, ** Receptive Fields by Self-Organization in Neural Networks,
S.Inawashiro, Y. Tamori and Y. Musya (Faculty of Engineering, Tohoku
University)

P1-7, ** Does Backpropagation Exist in Biological Systems? -- Discussions
and Considerations, Shyozo Yasui (Kyusyu Institute of Technology), Eiki
Hida (Tamagawa University)

P1-8, Accelerating the convergence of the error back-propagation
algorithm by deciding effective teacher signal, Yutaka Fukuoka, Hideo
Matsuki, Hidetake Muraoka and Haruyuki Minamitani (Keio Univ.)

P1-9, **Three-Layered Backpropagation Model with Temperature Parameter,
Yoji Fukuda, Manabu Kotani and Haruya Matsumoto (Faculty of Engineering,
Kobe University)

P1-10, Kalman Type Least Square Error Learning Law for Sequential Neural
Network and its Information Theory, Kazuyoshi Matsumoto (Kansai Advanced
Research Center, CRL, MPT)

P1-11, Learning Surface of Hierarchical Neural Networks and Valley
Learning Method, Kazutoshi Gouhara, Norifusa Kanai, Takeshi Iwata and
Yoshiki Uchikawa (Department of Electronic Mechanical Engineering School
of Engineering: Nagoya Univ.)
 
P1-12, A Study of Generalization of Multi-layered Neural Networks,
Katsumasa Matsuura (Mechanical Engineering Research Laboratory, Hitachi
Ltd.)

P1-13, When "Learning" occurs in Machine Learning, Noboru Watanabe
(Department of Biophysics, Kyoto University)

P1-14, A Study on Self-supervised Learning System, Kazushige Saga, Tamami
Sugasaka and Shigemi Nagata (Computer-Based Systems Laboratory Fujitsu
Laboratories Ltd.)

P1-15, A Learning Circuit for VLSI Analog Neural Network Implementation,
Hiroyuki Wasaki, Yoshihiko Horio and Shogo Nakamura (Department of
Electronic Engineering : Tokyo Denki Univ.)

P1-16, Comparison of Learning Methods for Recurrent Neural Networks,
Tatsumi Watanabe, Kazutoshi Gouhara and Yoshiki Uchikawa (Department of
Electronic Mechanical Engineering, School of Engineering : Nagoya Univ.)

P1-17, A Learning algorithm for the neural network of Hopfield type,
Fumio Matsunari and Masuji Ohshima (Toyota Central Res. & Develop. Labs.
Inc.)

P1-18, Quantitative relationship between internal model of motor system
and movement accuracy, movement distance and movement time, Yoji Uno and
Ryoji Suzuki (Faculty of Engineering: University of Tokyo)

P1-19, Autonomic control of bipedal locomotion using neural oscillators,
Gentaro Taga, Yoko Yamaguchi and Hiroshi Shimizu (Faculty of
Pharmaceutical Sciences: University of Tokyo)

P1-20, Learning Model of Posture Control in Cerebellum, Hiroaki Gomi and
Mitsuo Kawato (ATR Auditory and Visual Perception Research Laboratories)

P1-21, Hybrid Control of Robotic Manipulator by Neural Network Model
(Sensing and Hybrid Control of Robotic Manipulator with Collision
Phenomena), Takanori Shibata, Toshio Fukuda, Fumihito Arai and Hiroshi
Wada (Nagoya University), Masatoshi Tokita and Toyokazu Mituoka (Kisarazu
Technical College) Yasumasa Shoji (Toyo Engineering Corp.)

P1-22, Hybrid Position/Force Control of Robotic Manipulator by
Application of Neural Network (Adaptive Control with Consideration of
Characteristics of Objects), Masatoshi Tokita and Toyokazu
Mituoka(Kisarazu National College of Technology), Toshio Fukuda and
Takanori Shibata (Nagoya Univ.)

P1-23, Reverbration in Chaotic Neural Network with a Fractal Connection,
Masatoshi Hori, Masaaki Okabe and Masahiro Nakagawa (Department of
Electrical Engineering, Faculty of Engineering, Nagaoka University of
Technology)

P1-24, An investigation of correlation possibility in neurocomputing and
quantum mechanics via Matrix Dynamics, Tomoyuki Nishio (Technology
Planning office General R&D Laboratory JUKI Corporation)

P1-25, Knowledge Representation and Parallel Inference with Structured
Networks, Akira Namatame and Youichi Ousawa (National Defense Academy)

P1-26, Simulation of a Spiking Neuron with Multiple Input, G. Bugmann
(Fundamental Research Laboratories, NEC Corporation)

P2-1, Parallel Implementation of Edge detection by Energy Minimization on
QCDPAX machine, Hideki Asoh (Electrotechnical Laboratory), Youichi
Hachikubo and Tsutomu Hoshino (University of Tsukuba)

P2-2, Characteristics of the Marr-Hildreth filter for two particle image,
Shigeharu Toyoda (Department of Chemical Engineering, Faculty of
Engineering Science, Osaka Univ.)

P2-3, A neural network for fixation point selection, Makoto Hirahara and
Takashi Nagano (College of engineering, Hosei Univ.)

P2-4, A Method for Analyzing Inverse Dynamics of the Retinal Horizontal
Cell Response through the Ionic Current Model, Yoshimi Kamiyama, Hiroyuki
Ishii and Shiro Usui (Information and Computer Science, Toyohashi Uni. of
Technology)

P2-5, Selective Recognition of Plural Patterns by Neural Networks,
Katsuji Imai, Kazutoshi Gouhara and Yoshiki Uchikawa (Department of
Electronic Mechanical Engineering, School of Engineering, Nagoya Univ.)

P2-6, A neural network for size-invariant pattern recognition, Masaichi
Ishikawa and Toshi Nagano (College of Engineering, Hosei Univ.)

P2-7, Human-Face Identification by Neural Network for Mosaic Pattern,
Makoto Kosugi (NTT Human Interface Laboratories)

P2-8, Pattern Classification and Recognition Neural Network
based on the Associative Learning Rules, Hidetoshi Ikeno
(Maizuru College of Technology), Shiro Usui and Manabu
Sakakibara (Toyohashi Univ. of Technology)

P2-9, Learning of A Three-Layer Neural Network with Translated
Patterns, Jianqiang YI, Shuichi Kurogi and Kiyotoshi Matsuoka
(Division of Control Engineering, Kyushu Institute of 
Technology)

P2-10, Combining a Neural Network with Template Matching for Handwritten
Numeral Classification, Masahiko Tateishi, Haruaki Yamazaki (Systems
Laboratories, OKI Electric Corporation)

P2-11, Recognition of Continuous Writing of English Words with the
Mechanism of Selective Attention, Taro Imagawa and Kunihiko Fukushima
(Faculty of Engineering Science, Osaka University)

P2-12, Recognition of Handwritten Alphanumeric Characters by the
Neocognitron, Nobuaki Wake Kunihiko Fukushima (Faculty of Engineering
Science, Osaka University)

P2-13, Target Recognition with Chebyshev Networks, Nobuhisa Ueda and
Akira Namatame (National Defense Academy)

P2-14, A Large Scale Neural Network "Comb NET" for Printed Kanji
Character Recognition (JIS 1st and 2nd Level Character Set), Takashi
Tohma, Akira Iwata, Hiroshi Matsuo and Nobuo Suzumura (Nagoya Institute
of Technology)

P2-15, Discriminative Properties of the Temporal Pattern in the
Mesencephalic Periaqueductal Gray of Rat, Mitsuo Terasawa, Minoru
Tsukada, Makoto Mizuno, Takeshi Aihara (Faculty of Engineering Tamagawa
Univ.)

P2-16, Spoken Word Recognition using sequential Neural Network, Seiichi
Nakagawa and Isao Hayakawa (Toyohashi University of Technology, Dept. of
Info. & Comp. Science.)

P2-17, Learning of the three-vowel sequence by a neural network model and
influence to a middle vowel from preceding and succeeding vowels,
Teruhiko Ohtomo and Ken-ichi Hara (Faculty of Engineering, Yamagata
Univ.)

P2-18, A Self-Organizing Neural Network for The Classification of
Temporal Sequences, Seiichi Ozawa (Nara National College of Technology),
Kazuyoshi Tsutsumi (Faculty of Science and Technology, Ryukoku Univ.),
Haruya Matsumoto (Faculty of Technology, Kobe Univ.)

P2-19, Neural Mechanism for Fine Time-Resolution, Kiyohiko Nakamura
(Graduate School of Science and Engineering Tokyo Institute of
Technology)

P2-20, Neuromagnetic Image Reconstruction by a Neural Network, Hisashi
Tsuruoka (Fukuoka Institute of Technology) and Kohyu Fukunishi (Advanced
Research Lab., Hitachi, Ltd.)

P2-21, A Sparse Stabilization in Mutually Connected Neural Networks,
Hiroshi Yamakawa (Faculty of Engineering, University of Tokyo), Yoichi
Okabe (Research Center for Advanced Science and Technology, University of
Tokyo)

P2-22, ** Relation between Recollection Ability and Membrane Potential
Threshold in Hopfield Model, Eizo Ohno and Atsushi Yamanaka (Sharp
Laboratory)

P2-23, Weight quantization in learning Boltzmann machine, Masanobu
Takahashi, Wolfgang Balzer, Jun Ohta and Kazuo Kyuma (Solid state quantum
electronics department Central Research Laboratory Mitsubishi Electric
Corporation)

P2-24, A cellular organizing approach for the travelling salesman
problem, M. Shigematsu (Electrotechnical Laboratory)

P2-25, Combinatorial Optimization Problems and Stochastic Logic Neural
Network, Yoshikazu Kondo and Yasuji Sawada (Research Institute of
Electrical Communication Tohoku Univ.)

P2-26, Neural Network Models on SIMD Architecture Computer, Makoto
Hirayama (ATR Auditory and Visual Perception Research Laboratories)

P2-27, Optimal connection of an associative network with non-uniform
elements, Hiro-fumi Yanai (Department of Mathematical Engineering and
Information Physics, University of Tokyo)




------------------------------

Subject: Please consider this tutorial announcement for posting
From:    Ron Riesenbach <itrctor@csri.toronto.edu>
Date:    Tue, 06 Nov 90 17:24:08 -0500

Sir:

  The Information Technology Research Centre (ITRC) is a non-profit Centre
of Excellence of the province of Ontario.  We sponsor world-class research
in various areas of information technology and transfer the results of this
research to companies.

  The tutorial (announcement below) is given annually by Dr. Hinton and
has in the past attracted people from across North America.  I would
appreciate it if you would consider posting it in your neuron-digest.

  Thank you.

                                      Ron Riesenbach, Manager
                                      Information Technology Research Centre

 ----------------------------------------------------------------------
                       Neural Networks for Industry


                   from fundamentals to current research



                               presented by:
                            Dr. Geoffrey Hinton


                               Sponsored by:
                  Information Technology Research Centre
                                    and
                          PRECARN Associates Inc.



                         December 11 and 12, 1990
                         Regal Constellation Hotel
                               900 Dixon Rd.
                   (near the Lester B. Pearson Airport)
                             Toronto, Ontario





1.  Why Neural Networks?

     Serial computation has been very  successful  at  tasks  that  can  be
characterized  by clean logical rules.  It has been much less successful at
tasks like real-world perception or common sense  reasoning.   These  tasks
typically  require  massive amounts of uncertain evidence to be combined in
complex ways to reach reliable decisions.  The brain is extremely  good  at
these  computations  and  there  is  now a growing consensus that massively
parallel "neural" computation may be the best way to solve these problems.

     The resurgence of interest in neural  networks  has  been  fuelled  by
several factors. Powerful new search techniques such as simulated annealing
and its deterministic approximations can  be  embodied  very  naturally  in
these  networks.   As such, parallel hardware implementations promise to be
extremely fast at performing the best-fit searches  required  for  content-
addressable  memory  and real-world perception. Recently, new learning pro-
cedures have been developed which allow networks to  learn  from  examples.
The  learning  procedures  automatically construct the internal representa-
tions that the networks require in particular  domains,  and  so  they  may
remove  the need for explicit programming in ill-structured tasks that con-
tain a mixture of regular structure, partial regularities and exceptions.


2.  Who Should Attend?

     Day 1 is a tutorial directed at Industry Researchers and Managers  who
would  like  to understand the basic principles underlying neural networks.
The tutorial will explain the main learning procedures and show  how  these
are  used  effectively  in  current  applications.  No previous exposure to
neural networks is necessary although  a  degree  in  computer  science  or
electrical engineering (or equivalent experience) is desirable.

     Day 2 is an overview of advances made in this field in the last two or
three years. Research in progress at various laboratories will be reviewed.
This overview of how recent developments may lead to better learning proce-
dures in the future  will be best  appreciated by industry  researchers and
managers  who  have  some experience in this field or who have attended the
tutorial the previous day.

     Those attending both days can expect to gain an understanding  of  the
current  state-of-the-art  in  neural networks and be in a position to make
informed decisions about whether this technology is  currently  applicable,
or  may  soon  become  applicable,  to  specific  problems in their area of
interest.




                                   DAY 1


INTRODUCTION
     o Computers versus brains
     o The hardware of the brain
     o Cooperative computation
     o The Least Mean Squares learning procedure
     o The perceptron paradigm
     o Why hidden units are needed
     o Varieties of learning procedure
     o Competitive learning
     o Learning topographic maps


BACKPROPAGATION LEARNING AND SIMPLE APPLICATIONS
     o The backpropagation algorithm
     o The NetTalk example
     o The family trees example
     o The parity example
     o Theoretical results on generalization
     o Simplicity and generalization
     o Detecting bombs in suitcases
     o Following a road with an autonomous land vehicle


BACKPROPAGATION: COMPLEX APPLICATIONS AND VARIATIONS
     o Recognizing phonemes in spectrograms
     o Recognizing hand-printed digits
     o Alternative error functions
     o Converting hand movements into speech
     o Medical diagnosis
     o What makes an application feasible
     o The speed of convergence and ways to improve it
     o Coding the input for backpropagation
     o Self-supervised backpropagation


HOPFIELD NETS, BOLTZMANN MACHINES, AND MEAN FIELD NETS
     o Binary Hopfield nets and their limitations
     o Simulated annealing for escaping local energy minima
     o Boltzmann Machines
     o The Boltzmann machine learning procedure and its limitations
     o Mean field networks for faster search
     o Application to the travelling salesman problem
     o A learning procedure for mean field nets.




                                     DAY 2


RADIAL BASIS FUNCTIONS AND COMMUNITIES OF LOCAL EXPERTS
     o Radial Basis Functions
     o Relation to kernel methods in statistics
     o Relation to Kanerva memories
     o Application to predicting chaotic series
     o Application to shape recognition
     o Using soft competitive learning to adapt radial basis functions
     o The elastic net
     o Relation to mixtures of gaussians
     o Communities of expert networks
     o Relation to mixture models
     o Applications to vowel recognition


MORE UNSUPERVISED LEARNING METHODS
     o Finding multimodal projections of high dimensional data
     o Application to adaptive modems
     o Application to discovering important features of vowels
     o Preserving information about the input with limited channel capacity
     o Using coherence assumptions to discover spatial or temporal invariants
     o Applications to stereo fusion and shape recognition
     o Implementation in a new type of Boltzmann machine


NETWORKS FOR MODELLING SEQUENCES
     o Backpropagation in recurrent networks
     o Recurrent networks for predicting the next term in a sequence
     o Using predictive networks for data-compression
     o Ways to restrict recurrent networks
     o Applications of recurrent networks to sentence understanding
     o Hidden Markov Models and the Baum-Welch training procedure
     o Combining HMM's with feedforward networks
     o Implementing HMM recognizers in feedforward networks
     o Reinforcement learning and the temporal credit assignment problem
     o Recent developments in learning good action sequences


MISCELLANEOUS RECENT DEVELOPMENTS
     o Neural networks for solving very large optimization problems
     o Neural networks in non-linear controllers
     o Better networks for hand-printed character recognition
     o Why local minima are not fatal for backpropagation
     o Why the use of a validation set improves generalization
     o Adding hidden units incrementally
     o Polynomial nets





3.  Seminar Schedule

 Tuesday, December 11, 1990             Wednesday, December 12, 1990


 8:00-9:00   Registration and Coffee     8:00-9:00   Registration and Coffee

 9:00-9:05   Opening words               9:00-9:05   Opening words

 9:05-10:30  Tutorial Session #1         9:05-10:30  Advanced Session #1

10:30-11:00  Break                      10:30-11:00  Break

11:00-12:30  Tutorial Session #2         1:00-12:30  Advanced Session #2

12:30-2:00   Lunch                      12:30-2:00   Lunch

 2:00-3:30   Tutorial Session  #3        2:00-3:30   Advanced Session  #3

 3:30-4:00   Break                       3:30-4:00   Break

 4:00-5:30   Tutorial Session #4         4:00-5:30   Advanced Session #4

 5:30-6:30   Wine and Cheese reception   5:30        Closing Words


4.  Registration and Fees:

     Fees are based on the affiliation of  attendees.   Employees  of  com-
panies  who  are  members  of ITRC's Industrial Affiliates Program or whose
companies are members of  PRECARN pay a subsidized fee  of  $75/day.   Non-
members  fees  are $250/day.  There is no discount for attending both days.
Payment should be made  by  cheque  (Payable  to:  "Information  Technology
Research Centre") and should accompany the registration form where possible.
Due to limited space ITRC and Precarn members will have priority in case of
over-subscription.   ITRC and Precarn reserve the right to limit the number
of registrants from any one company.

     Fees include a copy of the course notes and transparencies, coffee and
light  refreshments at the breaks, a luncheon each day as well as an infor-
mal wine and cheese reception  Tuesday  evening  (open  to  registrants  of
either  day).   Participants are responsible for their own hotel accommoda-
tion, reservations and costs, including hotel breakfast, evening meals  and
transportation. PLEASE MAKE YOUR HOTEL RESERVATIONS EARLY:

                         Regal Constellation Hotel
                              900 Dixon Road
                            Etobicoke, Ontario
                                  M9W 1J7
                         Telephone: (416) 675-1500
                             Telex: 06-989511
                            Fax: (416) 675-1737

     When reserving hotel accommodation  please  mention  ITRC/PRECARN  for
special room rates.

     Registrations will be accepted up to 5:00pm December  6,  1990.   Late
registrations  may  be  accepted  but,  due to limited space, attendees who
register by December 6th will have priority over late registrants.

     To register, complete the registration form beolow then mail or fax it
to either one of the following two offices:

 ITRC, Rosanna Reid             PRECARN, Charlene Ferguson
 University of Toronto          30 Colonnade Rd., Suite 300
 D.L. Pratt Bldg., Rm. 286      Nepean, Ontario K2E 7J6
 Toronto, Ontario M5S 1A1       Phone: (613) 727-9576
 Phone: (416) 978-8558          Fax: (613) 727-5672
 Fax: (416) 978-7207





5.  Biography

Dr. Geoffrey E. Hinton (instructor)

     Geoffrey Hinton is Professor of Computer Science at the University  of
Toronto,  a Fellow of the Canadian Institute for Advanced Research, a prin-
cipal researcher with the Information Technology Research Centre and a pro-
ject leader with the Institute for Robotics and Intelligent Systems (IRIS).
He received his PhD in Artificial Intelligence from the University of Edin-
burgh.   He has been working on computational models of neural networks for
the last fifteen years and has published 70 papers  and  book  chapters  on
applications   of  neural  networks  in  vision,  learning,  and  knowledge
representation.  These publications include the book  "Parallel  Models  of
Associative  Memory"  (with James Anderson) and the original papers on dis-
tributed representations, on Boltzmann machines (with Terrence  Sejnowski),
and  on back-propagation (with David Rumelhart and Ronald Williams).  He is
also one of the major contributors to the recent collection "Parallel  Dis-
tributed Processing" edited by Rumelhart and McClelland.

     Dr. Hinton was formerly an Associate Professor of Computer Science  at
Carnegie-Mellon  University  where  he  created  the connectionist research
group and was responsible for the graduate course on "Connectionist Artifi-
cial  Intelligence".  He is on the governing board of the Cognitive Science
Society and the governing council of the American Association  for  Artifi-
cial  Intelligence.  He is a member of the editorial boards of the journals
Artificial Intelligence, Machine Learning, Cognitive Science, Neural Compu-
tation and Computer Speech and Language.

     Dr. Hinton is an expert at explaining neural  network  research  to  a
wide  variety  of audiences.  He has given invited lectures on the research
at numerous international conferences, workshops, and  summer  schools.  He
has  given  industrial  tutorials  in  the United States for the Technology
Transfer Institute, AT&T Bell labs, Apple,  Digital  Equipment  Corp.,  and
the American Association for Artificial Intelligence.





     -------------------------Registration Form -------------------

                       Neural Networks for Industry
                        Tutorial by Geoffrey Hinton
                           December 11-12, 1990
                    Regal Constellation, 900 Dixon Rd.


Name      _________________________________________
Title     _________________________________________
Organization  _____________________________________
Address   _________________________________________
          _________________________________________
          _________________________________________
Postal Code   ___________
Telephone  __________________ Fax _________________


Registration Fee (check those that apply):

                          Day 1          Day 2         Total
                         -------        -------       -------
ITRC or PRECARN member  __  $75        __  $75        ________
Non-member              __  $250       __  $250       ________


Please make cheques payable to Information Technology Research Centre
                  REGISTRATION DEADLINE:  DECEMBER 6/90.
                   (Space is limited so register early)

Please fax or mail your registration to ITRC or PRECARN:

  ITRC, Rosanna Reid           PRECARN, Charlene Ferguson
  University of Toronto        30 Colonnade Rd., Suite 300
  D.L. Pratt Bldg., Rm. 286    Nepean, Ontario
  6 King's College Rd.         K2E 7J6
  Toronto, Ontario M5S 1A1     phone (613) 727-9576
  phone (416) 978-8558         fax    (613) 727-5672
  fax    (416) 978-7207

     --------------------------------------------------------------


------------------------------

End of Neuron Digest [Volume 6 Issue 65]
****************************************