neuron-request@HPLABS.HP.COM ("Neuron-Digest Moderator Peter Marvit") (10/29/89)
Neuron Digest Saturday, 28 Oct 1989 Volume 5 : Issue 42 Today's Topics: NIPS'89 Postconference Workshops IJCNN 1990 - Request for Volunteers NIPS-89 workshop Send submissions, questions, address maintenance and requests for old issues to "neuron-request@hplabs.hp.com" or "{any backbone,uunet}!hplabs!neuron-request" Use "ftp" to get old issues from hplpm.hpl.hp.com (15.255.176.205). ------------------------------------------------------------ Subject: NIPS'89 Postconference Workshops From: Alex.Waibel@SPEECH2.CS.CMU.EDU Date: Fri, 06 Oct 89 21:19:38 -0400 Below are the preliminary program and brief descriptions of the workshop topics covered during this years NIPS-Postconference Workshops to be held in Keystone from November 30 through December 2 (right following the NIPS conference). Please register for both conference and Workshops using the general NIPS conference registration forms. With it, please indicate which workshop topic below you may be most interested in attending. Your preferences are in no way binding or limiting you to any particular workshop but will help us in allocating suitable meeting rooms and scheduling workshop sessions in an optimal way. For your convenience, you may simply include a copy of the form below with your registration material marking it for your three most prefered workshop choices in order of preference (1,2 and 3). For registration information (both NIPS conference as well as Postconference Workshops), please contact the Local Arrangements Chair, Kathie Hibbard, by sending email to hibbard@boulder.colorado.edu, or by writing to: Kathie Hibbard NIPS '89 University of Colorado Campus Box 425 Boulder, Colorado 80309-0425 For technical questions relating to individual conference workshops, please contact the individual workshop leaders listed below. Please feel free to contact me with any questions you may have about the workshops in general. See you in Denver/Keystone, Alex Waibel NIPS Workshop Program Chairman School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 412-268-7676, waibel@cs.cmu.edu ================================================================ ____________________________________________________________ ! POST CONFERENCE WORKSHOPS AT KEYSTONE ! ! THURSDAY, NOVEMBER 30 - SATURDAY, DECEMBER 2, 1989 ! !____________________________________________________________! Thursday, November 30, 1989 5:00 PM: Registration and Reception at Keystone Friday, December 1, 1989 7:30 - 9:30 AM: Small Group Workshops 4:30 - 6:30 PM: Small Group Workshops 7:30 - 10:30 PM: Banquet and Plenary Discussion Saturday, December 2, 1989 7:30 - 9:30 AM: Small Group Workshops 4:30 - 6:30 PM: Small Group Workshops 6:30 - 7:15 PM: Plenary Discussion, Summaries 7:30 - 11:00 PM: Fondue Dinner, MountainTop Restaurant ================================================================ PLEASE MARK YOUR PREFERENCES (1,2,3) AND ENCLOSE WITH REGISTRATION MATERIAL: - ----------------------------------------------------------------------------- ______1. LEARNING THEORY: STATISTICAL ANALYSIS OR VC DIMENSION? ______2. STATISTICAL INFERENCE IN NEURAL NETWORK MODELLING ______3. NEURAL NETWORKS AND GENETIC ALGORITHMS ______4. VLSI NEURAL NETWORKS: CURRENT MILESTONES AND FUTURE HORIZONS ______5. APPLICATION OF NEURAL NETWORK PROCESSING TECHNIQUES TO REAL WORLD MACHINE VISION PROBLEMS ______6. IMPLEMENTATIONS OF NEURAL NETWORKS ON DIGITAL, MASSIVELY PARALLEL COMPUTERS ______7. LARGE, FAST, INTEGRATED SYSTEMS BASED ON ASSOCIATIVE MEMORIES ______8. NEURAL NETWORKS FOR SEQUENTIAL PROCESSING WITH APPLICATIONS IN SPEECH RECOGNITION ______9. LEARNING FROM NEURONS THAT LEARN ______10. NEURAL NETWORKS AND OPTIMIZATION PROBLEMS 11. (withdrawn) ______12. NETWORK DYNAMICS ______13. ARE REAL NEURONS HIGHER ORDER NETS? ______14. NEURAL NETWORK LEARNING: MOVING FROM BLACK ART TO A PRACTICAL TECHNOLOGY ______15. OTHERS ?? __________________________________________________ 1. LEARNING THEORY: STATISTICAL ANALYSIS OR VC DIMENSION? Sara A. Solla AT&T Bell Laboratories Crawford Corner Road Holmdel, NJ 07733-1988 Phone: (201) 949-6057 E-mail: solla@homxb.att.com Recent success at describing the process of learning in layered neural networks and the resulting generalization ability has emerged from two different approaches. Work based on the concept of VC dimension emphasizes the connection between learning and statistical inference in order to analyze questions of bias and variance. The statistical approach uses an ensemble description to focus on the prediction of network performance for a specific task. Participants interested in learning theory are invited to discuss the differences and similarities between the two approaches, the mathematical relation between them, and their respective range of applicability. Specific questions to be discussed include comparison of predictions for required training set sizes, for the distribution of generalization abilities, for the probability of obtaining good performance with a training set of fixed size, and for estimates of problem complexity applicable to the determination of learning times. 2. STATISTICAL INFERENCE IN NEURAL NETWORK MODELLING Workshop Chair: Richard Golden Stanford University Psychology Department Stanford, CA 94305 (415) 725-2456 E-mail: golden@psych.stanford.edu This workshop is designed to show how the theory of statistical inference is directly applicable to some difficult neural network modelling problems. The format will be tutorial in nature (85% informal lecture, 15% discussion). Topics to be discussed include: obtaining probability distributions for neural networks, interpretation and derivation of optimal learning cost functions, evaluating the generalization performance of networks, asymptotic sampling distributions of network weights, statistical mechanics calculation of learning curves in some simple examples, statistical tests for comparing internal representations and deciding which input units are relevant to the prediction task. Dr. Naftali Tishby (AT&T Bell Labs) and Professor Halbert White (UCSD Economics Department) are the invited experts. 3. Title: NEURAL NETWORKS AND GENETIC ALGORITHMS Organizers: Lawrence Davis (Bolt Beranek and Newman, Inc.) Michael Rudnick (Oregon Graduate Center) Description: Genetic algorithms have many interesting relationships with neural networks. Recently, a number of researchers have investigated some of these relationships. This workshop will be the first forum bringing those researchers together to discuss the current and future directions of their work. The workshop will last one day and will have three parts. First, a tutorial on genetic algorithms will be given, to ground those unfamiliar with the technology. Second, seven researchers will summarize their results. Finally there will be an open discussion on the topics raised in the workshop. We expect that anyone familiar with neural network technology will be comfortable with the content and level of discussion in this workshop. 4. VLSI NEURAL NETWORKS: CURRENT MILESTONES AND FUTURE HORIZONS Moderators: Joshua Alspector and Daniel B. Schwartz Bell Communications Research GTE Laboratories, Inc. 445 South Street 40 Sylvan Road Morristown, NJ 07960-19910 Waltham, MA 02254 (201) 829-4342 (617) 466-2414 e-mail: josh@bellcore.com e-mail: dbs%gte.com@relay.cs.net This workshop will explore the areas of applicability of neural network implementations in VLSI. Several speakers will discuss their present implementations and speculate about where their work may lead. Workshop attendees will then be encouraged to organize working groups to address several issues which will be raised in connection with the presentations. Although it is difficult to predict which issues will be selected, some examples might be: 1) Analog vs. digital implementations. 2) Limits to VLSI complexity for neural networks. 3) Algorithms suitable for VLSI architectures. The working groups will then report results which will be included in the workshop summary. 5. APPLICATION OF NEURAL NETWORK PROCESSING TECHNIQUES TO REAL WORLD MACHINE VISION PROBLEMS Paul J. Kolodzy (617) 981-3822 kolodzy@ll.ll.mit.edu Murali M. Menon (617) 981-5374 This workshop will discuss the application of neural networks to vision applications, including image restoration and pattern recognition. Participants will be asked to present their specific application for discussion to highlight the relevant issues. Examples of such issues include, but are not limited to, the use of deterministic versus stochastic search procedures for neural network processing, using networks to extract shape, scale and texture information for recognition and using network mapping techniques to increase data separability. The discussions will be driven by actual applications with an emphasis on the advantages of using neural networks at the system level in addition to the individual processing steps. The workshop will attempt to cover a wide breadth of network architectures and invites participation from researchers in machine vision, neural network modeling, pattern recognition and biological vision. 6. IMPLEMENTATIONS OF NEURAL NETWORKS ON DIGITAL, MASSIVELY PARALLEL COMPUTERS Dr. K. Wojtek Przytula and Prof. S.Y. Kung Hughes Research Laboratories, RL 69 3011 Malibu Cyn. Road Malibu, CA 90265 Phone: (213) 317-5892 E-mail: wojtek%csfvax@hac2arpa.hac.com Implementations of neural networks span a full spectrum from software realizations on general-purpose computers to strictly special-purpose hardware realizations. Implementations on programmable, parallel machines, which are to be discussed during the workshop, constitute a compromise between the two extremes. The architectures of programmable parallel machines reflect the structure of neural network models better than those of sequential machines, thus resulting in higher processing speed. The programmability provides more flexibility than is available in specialized hardware implementations and opens a way for realization of various models on a single machine. The issues to be discussed include: mapping neural network models onto existing parallel machines, design of specialized programmable parallel machines for neural networks, evaluation of performance of parallel machines for neural networks, uniform characterization of the computational requirements of various neural network models from the point of view of parallel implementations. 7. LARGE, FAST, INTEGRATED SYSTEMS BASED ON ASSOCIATIVE MEMORIES Michael R. Raugh Director of Learning Systems Division Research Institute for Advanced Computer Science (RIACS) NASA Ames Research Center, MS 230-5 Moffett Field, CA 94035 e-mail: raugh@riacs.edu Phone: (415) 694-4998 This workshop will address issues in the construction of large systems that have thousands or even millions of hidden units. It will present and discuss alternatives to backpropagation that allow large systems to learn rapidly. Examples from image analysis, weather prediction, and speech transcription will be discussed. The focus on backpropagation with its slow learning has kept researchers from considering such large systems. Sparse distributed memory and related associative-memory structures provide an alternative that can learn, interpolate, and abstract, and can do so rapidly. The workshop is open to everyone, with special encouragement to those working in learning, time-dependent networks, and generalization. 8. NEURAL NETWORKS FOR SEQUENTIAL PROCESSING WITH APPLICATIONS IN SPEECH RECOGNITION Herve Bourlard Philips Research Laboratory Brussels Av. Van Becelaere 2, Box 8 B-1170 Brussels, Belgium Phone: 011-32-2-674-22-74 e-mail address: bourlard@prlb.philips.be or: prlb2!bourlard@uunet.uu.net Speech recognition must contend with the statistical and sequential nature of the human speech production system. Hidden Markov Models (HMM) provide a powerful method to cope with both of these, and their use made a breakthrough in speech recognition. On the other hand, neural networks have recently been recognized as an alternative tool for pattern recognition problems such as speech recognition. Their main useful properties are their discriminative power and their capability to deal with non-explicit knowledge. However, the sequential aspect remains difficult to handle in connectionist models. If connections are supplied with delays, feedback loops can be added providing dynamic and implicit memory. However, in the framework of continuous speech recognition, it is still difficult to use only neural networks for the segmentation and recognition of a sentence into a sequence of speech units, which is efficiently solved in the HMM approach by the well known ``Dynamic Time Warping'' algorithm. This workshop should be the opportunity for reviewing neural network architectures which are potentially able to deal with sequential and stochastic inputs. It should also be discussed to which extent the different architectures can be useful in recognizing isolated units (phonemes, words, ...) or continuous speech. Amongst others, we should consider spatiotemporal models, time-delayed neural networks (Waibel, Sejnowsky), temporal flow models (Watrous), hidden-to-input (Elman) or output-to-input (Jordan) recurrent models, focused back-propagation networks (Mozer) or hybrid approaches mixing neural networks and standard sequence matching techniques (Sakoe, Bourlard). 9. LEARNING FROM NEURONS THAT LEARN Moderated by Thomas P. Vogl Environmental Research Institute of Michigan 1501 Wilson Blvd. Arlington, VA 22209 Phone: (703) 528-5250 E-mail: TVP%nihcu.bitnet@cunyvm.cuny.edu FAX: (703) 524-3527 In furthering our understanding of artificial and biological neural systems, the insights that can be gained from the perceptions of those trained in other disciplines can be particularly fruitful. Computer scientists, biophysicists, engineers, psychologists, physicists, and neurobiologists tend to have different perspectives and conceptions of the mechanisms and components of "neural networks" and to weigh differently their relative importance. The insights obvious to practitioners of one of these disciplines are often far from obvious to those trained in another, and therefore may be especially relevant to the solutions of ornery problems. The workshop provides a forum for the interdisciplinary discussion of biological and artificial networks and neurons and their behavior. Informal group discussion of ongoing research, novel ideas, approaches, comparisons, and the sharing of insights will be emphasized. The specific topics to be considered and the depth of the analysis/discussion devoted to any topic will be determined by the interest and enthusiasm of the participants as the discussion develops. Participants are encouraged to consider potential topics in advance, and to present them informally but succinctly (under five minutes) at the beginning of the workshop. 10. NEURAL NETWORKS AND OPTIMIZATION PROBLEMS ---------------------------------------- Prof. Carsten Peterson University of Lund Dept. of Theoretical Physics Solvegatan 14A S-223 62 Lund Sweden phone: 011-46-46-109002 bitnet: THEPCAP%SELDC52 Workshop description: The purpose of the workshop is twofold; to establish the present state of the art and to generate novel ideas. With respect to the former, firm answers to the following questions should emerge: (1). Does the Hopfield- Tank approach or variants thereof really work with respect to quality, reliability, parameter insensitivity and scalability? (2). If this is the case, how does it compare with other cellular approaches like "elastic snake" and genetic algorithms? Novel ideas should focus on new encoding schemes and new application areas (in particular, scheduling problems). Also, if time allows, optimization of neural network learning architectures will be covered. People interested in participating are encouraged to communicate their interests and expertise to the chairman via e-mail. This would facilitate the planning. 12. Title: NETWORK DYNAMICS Chair: Richard Rohwer Centre for Speech Technology Research Edinburgh University 80, South Bridge Edinburgh EH1 1HN, Scotland Phone: (44 or 0) (31) 225-8883 x280 e-mail: rr%uk.ac.ed.eusip@nsfnet-relay.ac.uk Summary: This workshop will be an attempt to gather and improve our knowledge about the time dimension of the activation patterns produced by real and model neural networks. This broad subject includes the description, interpretation and design of these temporal patterns. For example, methods from dynamical systems theory have been used to describe the dynamics of network models and real brains. The design problem is being approached using dynamical training algorithms. Perhaps the most important but least understood problems concern the cognitive and computational significance of these patterns. The workshop aims to summarize the methods and results of researchers from all relevant disciplines, and to draw on their diverse insights in order to frame incisive, approachable questions for future research into network dynamics. Richard Rohwer JANET: rr@uk.ac.ed.eusip Centre for Speech Technology Research ARPA: rr%uk.ac.ed.eusip@nsfnet-relay.ac.uk Edinburgh University BITNET: rr@eusip.ed.ac.uk, 80, South Bridge rr%eusip.ed.UKACRL Edinburgh EH1 1HN, Scotland UUCP: ...!{seismo,decvax,ihnp4} !mcvax!ukc!eusip!rr PHONE: (44 or 0) (31) 225-8883 x280 FAX: (44 or 0) (31) 226-2730 13. ARE REAL NEURONS HIGHER ORDER NETS? Most existing artificial neural networks have processing elements which are computationally much simpler than real neurons. One approach to enhancing the computational capacity of artificial neural networks is to simply scale up the number of processing elements, but there are limits to this. An alternative is to build modules or subnets and link these modules in a larger net. Several groups of investigators have begun to analyze the computational abilities of real single neurons in terms of equivalent neural nets, in particular higher order nets, in which the inputs explicitly interact (eg. sigma-pi units). This workshop would introduce participants to the results of these efforts, and examine the advantages and problems of applying these complex processors in larger networks. Dr. Thomas McKenna Office of Naval Research Div. Cognitive and Neural Sciences Code 1142 Biological Intelligence 800 N. Quincy St. Arlington, VA 22217-5000 phone:202-696-4503 email: mckenna@nprdc.arpa mckenna@nprdc.navy.mil 14. NEURAL NETWORK LEARNING: MOVING FROM BLACK ART TO A PRACTICAL TECHNOLOGY Scott E. Fahlman School of Computer Science Carnegie-Mellon University Pittsburgh, PA 15213 Internet: fahlman@cs.cmu.edu Phone: (412) 268-2575 There are a number of competing algorithms for neural network learning, all rather new and poorly understood. Where theory is lacking, a reliable technology can be built on shared experience, but it usually takes a long time for this experience to accumulate and propagate through the community. Currently, each research group has its own bag of tricks and its own body of folklore about how to attack certain kinds of learning tasks and how to diagnose the problem when things go wrong. Even when groups are willing to share their hard-won experience with others, this can be hard to accomplish. This workshop will bring together experienced users of back-propagation and other neural net learning algorithms, along with some interested novices, to compare views on questions like the following: I. Which algorithms and variations work best for various classes of problems? Can we come up with some diagnostic features that tell us what techniques to try? Can we predict how hard a given problem will be? II. Given a problem, how do we go about choosing the parameters for various algorithms? How do we choose what size and shape of network to try? If our first attempt fails, are there symptoms that can tell us what to try next? III. What can we do to bring more coherence into this body of folklore, and facilitate communication of this informal kind of knowledge? An online collection of standard benchmarks and public-domain programs is one idea, already implemented at CMU. How can we improve this, and what other ideas do we have? ------------------------------ Subject: IJCNN 1990 - Request for Volunteers From: Karen Haines <khaines@GALILEO.ECE.CMU.EDU> Date: Mon, 09 Oct 89 14:33:05 -0400 This is a first call for volunteers to help at the IJCNN conference, to be held at the Omni Shorham Hotel in Washington D.C., on January 15-19, 1990. Full admittance to the conference and a copy of the proceedings is offered in exchange for your assistance throughout the conference. In general, each volunteer is expected to work one shift, either in the morning or the afternnon, each day of the conference. Hours for morning shift are, approximately, 7:00 am until 12:00 noon, and for the afternoon, 12:00 noon to 5:00 pm. In addition, assistance will be required for the social events. If you can`t work all week long please contact Karen Haines to see what can be worked out. There will be a mandatory meeting for all volunteers on January 14. To sign up please contact: Karen Haines - Volunteer Coordinator 3138 Beechwood Blvd. Pittsburgh, PA 15217 office: (412) 268-3304 message: (412) 422-6026 email: khaines@galileo.ece.cmu.edu or, Nina Kowalski - Assistant Volunteer Coordinator 209 W. 29th St. FLR 2 Baltimore, MD 21211 message: (301) 889-0587 email: nina@alpha.ece.jhu.edu If you have further questions, please feel free to contact me. Thank you, Karen Haines ------------------------------ Subject: NIPS-89 workshop From: Scott.Fahlman@B.GP.CS.CMU.EDU Date: Wed, 18 Oct 89 11:18:58 -0400 The following workshop is one of those scheduled for December 1 and 2 in Keystone Colorado, immediately following the NIPS-89 conference. I am sending the announcement to this list because the issues to be discussed at the workshop ought to be of particular interest to readers of this list. It is my hope to get a bunch of experienced net-hackers together in one place and to compare notes about the practical issues that arise in applying this technology -- the sort of stuff that doesn't usually show up in published papers. In addition to backprop people, I hope to have some people in the workshop who have practical experience with various other learning algorithms. I'd like to get some idea of who might attend this workshop. Please send a note to me (sef@cs.cmu.edu -- NOT to nn-bench!) if you think you might be there. Please indicate whether you are probable or just possible, depending on what other workshops look interesting. Also please indicate whether you think you'll have some experiences of your own to share, or whether you're basically a spectator, hoping to pick up some tips from others. The more active participants we get, the more valuable this will be for everyone. This is a one-day workshop, so it can be combined with certain others. I've requested a slot on the first day, but that's not settled yet. If there's some other one-day post-NIPS workshop on the list that you'd really like to attend as well, please tell me and I'll pass a summary along to the people doing the scheduling. - -- Scott *************************************************************************** NEURAL NETWORK LEARNING: MOVING FROM BLACK ART TO A PRACTICAL TECHNOLOGY Scott E. Fahlman School of Computer Science Carnegie-Mellon University Pittsburgh, PA 15213 Internet: fahlman@cs.cmu.edu Phone: (412) 268-2575 There are a number of competing algorithms for neural network learning, all rather new and poorly understood. Where theory is lacking, a reliable technology can be built on shared experience, but it usually takes a long time for this experience to accumulate and propagate through the community. Currently, each research group has its own bag of tricks and its own body of folklore about how to attack certain kinds of learning tasks and how to diagnose the problem when things go wrong. Even when groups are willing to share their hard-won experience with others, this can be hard to accomplish. This workshop will bring together experienced users of back-propagation and other neural net learning algorithms, along with some interested novices, to compare views on questions like the following: 1. Which algorithms and variations work best for various classes of problems? Can we come up with some diagnostic features that tell us what techniques to try? Can we predict how hard a given problem will be? 2. Given a problem, how do we go about choosing the parameters for various algorithms? How do we choose what size and shape of network to try? If our first attempt fails, are there symptoms that can tell us what to try next? 3. What can we do to bring more coherence into this body of folklore, and facilitate communication of this informal kind of knowledge? An online collection of standard benchmarks and public-domain programs is one idea, already implemented at CMU. How can we improve this, and what other ideas do we have? ------------------------------ End of Neurons Digest *********************