neuron-request@HPLABS.HP.COM (Neuron-Digest Moderator Peter Marvit) (11/14/88)
Neuron Digest Sunday, 13 Nov 1988 Volume 4 : Issue 22 Today's Topics: Seminar: A Connectionst Framework for visual recognition Reprints avail. Congress on Cybernetics and Systems TR available Tech report available Paul Thagard to speak on Analogical thinking Schedule of remaining talks this semester Tech. Report available Tech report on connectionist knowledge processing Technical Report Available [[Editor's Note: This issue and the next will be devoted to the backlog of technical talks and papers. Apologies in advance for notice of past talks. The issue on Consciousness will also have to wait. :-( -PM ]] Send submissions, questions, address maintenance and requests for old issues to "neuron-request@hplabs.hp.com" or "{any backbone,uunet}!hplabs!neuron-request" ------------------------------------------------------------ Subject: Seminar: A Connectionst Framework for visual recognition From: pratt@paul.rutgers.edu (Lorien Y. Pratt) Date: Tue, 18 Oct 88 09:49:23 -0400 I saw this posted locally, thought you might like to attend. Don't forget Josh Alspector's talk this Friday (10/21) on his Boltzmann chip! Rutgers University CAIP (Center for Computer Aids for Industiral Productivity. (CAIP) Seminar: A Connectionist Framework for Visual Recognition Ruud Bolle Exploratory Computer Vision Group IBM Thomas J. Watson Research Center Abstract This talk will focus on the organization and implementation of a vision system to recognize 3D objects. The visual world being modeled is assumed to consist of objects that can be represented by planar patches, patches of quadrics of revolution, and the intersection curves of those quadric surfaces. A significant portion of man-made objects can be represented using such prmitives. One of the contributions of this vision system is that fundamentally different feature types, like survface and curve descriptions, and simultaneously extracted and combined to index into a database of objects. The input to the system is a depth map of a scene comprising of one or more objects. From the depth map, surface parameters and surface intersection/object limb parameters are extracted. Parameter extraction is modeled as a set of layered and concurrent parameter space transforms. Any one transform computes only a partial geometric description that forms the input to a next transform. The final transform is a mapping into an object database, which can be viewed as the highest-level of confidence for geomeetric descriptoins and 3D objects within the parameter spaces. The approach is motivated by connectionist model of visual recognition systems. Date: Friday, November 4, 1988 Time: 3:00 PM Place: Conference room 139, CAIP center, Busch Campus For information, call (201) 932-3443 ------------------------------ Subject: Reprints avail. From: gluck@psych.Stanford.EDU (Mark Gluck) Date: Mon, 24 Oct 88 10:07:57 -0700 Reprints of the following two papers are available by netrequest to gluck@psych.stanford.edu or by writing: Mark Gluck, Dept. of Psychology, Jordan Hall; Bldg. 420, Stanford Univ., Stanford, CA 94305. Gluck, M. A., & Bower, G. H. (1988) From conditioning to category learning: An adaptive network model. Journal of Experimental Psychology: General, V. 117, N. 3, 227-247 Abstract -------- We used adaptive network theory to extend the Rescorla-Wagner (1972) least mean squares (LMS) model of associative learning to phenomena of human learning and judgment. In three experiments subjects learned to categorize hypothetical patients with particular symptom patterns as having certain diseases. When one disease is far more likely than another, the model predicts that subjects will sub- stantially overestimate the diagnosticity of the more valid symptom for the rare disease. The results of Experiments 1 and 2 provide clear support for this prediction in contradistinction to predictions from probability matching, exemplar retrieval, or simple prototype learning models. Experiment 3 contrasted the adaptive network model with one predicting pattern-probability matching when patients always had four symptoms (chosen from four opponent pairs) rather than the presence or absence of each of four symptoms, as in Experiment 1. The results again support the Rescorla-Wagner LMS learning rule as embedded within an adaptive network. Gluck, M. A., Parker, D. B., & Reifsnider, E. (1988) Some biological implications of a differential-Hebbian learning rule. Psychobiology, Vol. 16(3), 298-302 Abstract -------- Klopf (1988) presents a formal real-time model of classical conditioning which generates a wide range of behavioral Pavlovian phenomena. We describe a replication of his simulation results and summarize some of the strengths and shortcomings of the drive- reinforcement model as a real-time behavioral model of classical conditioning. To facilitate further comparison of Klopf's model with neuronal capabilities, we present a pulse-coded reformulation of the model that is more stable and easier to compute than the original, frequency-based model. We then review three ancillary assumptions to the model's learning algorithm, noting that each can be seen as dually motivated by both behavioral and biological considerations. ------------------------------ Subject: Congress on Cybernetics and Systems From: SPNHC@CUNYVM.CUNY.EDU (Spyros Antoniou) Organization: The City University of New York - New York, NY Date: 28 Oct 88 04:15:46 +0000 WORLD ORGANIZATION OF SYSTEMS AND CYBERNETICS 8 T H I N T E R N A T I O N A L C O N G R E S S O F C Y B E R N E T I C S A N D S Y S T E M S JUNE 11-15, 1990 at Hunter College, City University of New York, USA This triennial conference is supported by many international groups concerned with management, the sciences, computers, and technology systems. The 1990 Congress is the eighth in a series, previous events having been held in London (1969), Oxford (1972), Bucharest (1975), Amsterdam (1978), Mexico City (1981), Paris (1984) and London (1987). The Congress will provide a forum for the presentation and discussion of current research. Several specialized sections will focus on computer science, artificial intelligence, cognitive science, biocybernetics, psychocybernetics and sociocybernetics. Suggestions for other relevant topics are welcome. Participants who wish to organize a symposium or a section, are requested to submit a proposal ( sponsor, subject, potential participants, very short abstracts ) as soon as possible, but not later than September 1989. All submissions and correspondence regarding this conference should be addressd to: Prof. Constantin V. Negoita Congress Chairman Department of Computer Science Hunter College City University of New York 695 Park Avenue, New York, N.Y. 10021 U.S.A. =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= | Spyros D. Antoniou SPNHC@CUNYVM.BITNET SDAHC@HUNTER.BITNET | | | | Hunter College of the City University of New York U.S.A. | =-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-= ------------------------------ Subject: TR available From: Mark.Derthick@MCC.COM Date: Fri, 28 Oct 88 13:27:00 -0500 For copies of my thesis, ask copetas@cs.cmu.edu for CMU-CS-88-182 "Mundane Reasoning by Parallel Constraint Satisfaction." I am 1200 miles away from the reports, so asking me doesn't do you any good: Mark Derthick MCC 3500 West Balcones Center Drive Austin, TX 78759 (512)338-3724 Derthick@MCC.COM If you have previously asked me for this report, it should be arriving soon. There aren't many extra copies right now, so requests to copetas may be delayed for a while. ABSTRACT: Connectionistq networks are well suited to everyday common sense reasoning. Their ability to simultaneously satisfy multiple soft constraints allows them to select from conflicting information in finding a plausible interpretation of a situation. However these networks are poor at reasoning using the standard semantics of classical logic, based on truth in all possible models. This thesis shows that using an alternate semantics, based on truth in a single most plausible model, there is an elegant mapping from theories expressed using the syntax of propositional logic onto connectionist networks. An extension of this mapping to allow for limited use of quantifiers suffices to build a network from knowledge bases expressed in a frame language similar to KL-ONE. Although finding optimal models of these theories is intractable, the networks admit a fast hill climbing search algorithm that can be tuned to give satisfactory answers in familiar situations. The Role Shift problem illustrates the potential of this approach to harmonize conflicting information, using structured distributed representations. Although this example works well, much remains before realistic domains are feasible. ------------------------------ Subject: Tech report available From: Tony Robinson <ajr@DSL.ENG.CAM.AC.UK> Date: Mon, 31 Oct 88 11:14:50 +0000 Here is the summary of a tech report which demonstates that the error propagation algorithm is not limited to weighted-sum type nodes, but can be used to train radial-basis-function type nodes and others. Send me some email if you would like a copy. Tony. `'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`'`' Generalising the Nodes of the Error Propagation Network CUED/F-INFENG/TR.25 A J Robinson, M Niranjan, F Fallside Cambridge University Engineering Department Trumpington Street, Cambridge, England email: ajr@uk.ac.cam.eng.dsl 1 November 1988 Gradient descent has been used with much success to train connectionist models in the form of the Error Propagation Network (Rumelhart Hinton and Williams, 1986). In these nets the output of a node is a non-linear function of the weighted sum of the activations of other nodes. This type of node defines a hyper-plane in the input space, but other types of nodes are possible. For example, the Kanerva Model (Kanerva 1984), the Modified Kanerva Model (Prager and Fallside 1988), networks of Spherical Graded Units (Hanson and Burr, 1987), networks of Localised Receptive Fields (Moody and Darken, 1988) and the method of Radial Basis Functions (Powell, 1985; Broomhead and Lowe 1988) all use nodes which define volumes in the input space. Niranjan and Fallside (1988) summarise these and compare the class boundaries formed by this family of networks with feed-forward networks and nearest neighbour classifiers. This report shows that the error propagation algorithm can be used to train general types of node. The example of a gaussian node is given and this is compared with other connectionist models for the problem of recognition of steady state vowels from multiple speakers. ------------------------------ Subject: Paul Thagard to speak on Analogical thinking From: pratt@paul.rutgers.edu (Lorien Y. Pratt) Date: Mon, 31 Oct 88 12:57:09 -0500 COGNITIVE PSYCHOLOGY FALL COLLOQUIUM SERIES (Rutgers University) Date: 9 November 1988 Time: 4:30 PM Place: Room 307, Psychology Building, Busch Campus Paul Thagard, Cognitive Science Program, Princeton University ANALOGICAL THINKING Analogy is currently a very active area of research in both cognitive psychology and artificial intelligence. Keith Holyoak and I have developed connectionist models of analogical retrieval and mapping that are consistent with the results of psychological experiments. The new models use localist networks to simultaneously satisfy a set of semantic, structural, and pragmatic constraints. After providing a general view of analogical thinking, this talk will describe our model of analog retrieval. ------------------------------ Subject: Schedule of remaining talks this semester From: pratt@paul.rutgers.edu (Lorien Y. Pratt) Date: Mon, 31 Oct 88 13:13:18 -0500 Speaker schedule as of 10/31/88 for end of the semester talks in the Fall, 1988 Neural Networks Colloquium Series at Rutgers. Speaker Date Title - ------- ---- ----- Jack Gelfand 11/4/88 Neural nets, Intelligent Machines, and the AI wall Mark Jones 11/11/88 Knowledge representation in connectionist networks, including inheritance reasoning and default logic. E. Tzanakou 11/18/88 ALOPEX: Another optimization method Stefan Shrier 12/9/88 Abduction Machines for Grammar Discovery ------------------------------ Subject: Tech. Report available From: Vijaykumar Gullapalli 545-1596 <VIJAYKUMAR@cs.umass.EDU> Date: Mon, 31 Oct 88 15:57:00 -0400 The following Tech. Report is available. Requests should be sent to "SMITH@cs.umass.edu". A Stochastic Algorithm for Learning Real-valued Functions via Reinforcement Feedback Vijaykumar Gullapalli COINS Technical Report 88-91 University of Massachusetts Amherst, MA 01003 ABSTRACT Reinforcement learning is the process by which the probability of the response of a system to a stimulus increases with reward and decreases with punishment. Most of the research in reinforcement learning (with the exception of the work in function optimization) has been on problems with discrete action spaces, in which the learning system chooses one of a finite number of possible actions. However, many control problems require the application of continuous control signals. In this paper, we present a stochastic reinforcement learning algorithm for learning functions with continuous outputs. Our algorithm is designed to be implemented as a unit in a connectionist network. We assume that the learning system computes its real-valued output as some function of a random activation generated using the Normal distribution. The activation at any time depends on the two parameters, the mean and the standard deviation, used in the Normal distribution, which, in turn, depend on the current inputs to the unit. Learning takes place by using our algorithm to adjust these two parameters so as to increase the probability of producing the optimal real value for each input pattern. The performance of the algorithm is studied by using it to learn tasks of varying levels of difficulty. Further, as an example of a potential application, we present a network incorporating these real-valued units that learns the inverse kinematic transform of a simulated 3 degree-of-freedom robot arm. ------------------------------ Subject: Tech report on connectionist knowledge processing From: Charles Dolan <cpd@CS.UCLA.EDU> Date: Mon, 31 Oct 88 13:28:05 -0800 Implementing a connectionist production system using tensor products September, 1988 UCLA-AI-88-15 CU-CS-411-88 Charles P. Dolan Paul Smolensky AI Center Department of Computer Science & Hughes Research Labs Institute of Cognitive Science 3011 Malibu Canyon Rd. University of Colorado Malibu, CA 90265 Boulder, CO 80309-0430 & UCLA AI Laboratory Abstract In this paper we show that the tensor product technique for constructing variable bindings and for representing symbolic structure-used by Dolan and Dyer (1987) in parts of a connectionist story understanding model, and analyzed in general terms in Smolensky (1987)-can be effectively used to build a simplified version of Touretzky & Hinton's (1988) Distributed Connectionist Production System. The new system is called the Tensor Product Product System (TPPS). Copyright c 1988 by Charles Dolan & Paul Smolensky. For copies send a message to valerie@cs.ucla.edu at UCLA or kate@boulder.colorado.edu Boulder ------------------------------ Subject: Technical Report Available From: Dr Michael G Dyer <dyer@CS.UCLA.EDU> Date: Tue, 01 Nov 88 11:24:52 -0800 Symbolic NeuroEngineering for Natural Language Processing: A Multilevel Research Approach. Michael G. Dyer Tech. Rep. UCLA-AI-88-14 Abstract: Natural language processing (NLP) research has been built on the assumption that natural language tasks, such as comprehension, generation, argumentation, acquisition, and question answering, are fundamentally symbolic in nature. Recently, an alternative, subsymbolic paradigm has arisen, inspired by neural mechanisms and based on parallel processing over distributed representations. In this paper, the assumptions of these two paradigms are compared and contrasted, resulting in the observation that each paradigm possesses strengths exactly where the other is weak, and vice versa. This observation serves as a strong motivation for synthesis. A multilevel research approach is proposed, involving the construction of hybrid models, to achieve the long-term goal of mapping high-level cognitive function into neural mechanisms and brain architecture. Four levels of modeling are discussed: knowledge engineering level, localist connectionist level, distributed processing level, and artificial neural systems dynamics level. The two major goals of research at each level are (a) to explore its scope and limits and (b) to find mappings to the levels above and below it. In this paper the capabilities of several NLP models, at each level, are described, along with major research questions remaining to be resolved and major techniques currently being used in an attempt to complete the mappings. Techniques include: (1) forming hybrid systems with spreading activation, thresholds and markers to propagate bindings, (2) using extended back-error propagation in reactive training environments to eliminate microfeature representations, (3) transforming weight matrices into patterns of activation to create virtual semantic networks, (4) using conjunctive codings to implement role bindings, and (5) employing firing patterns and time-varying action potential to represent and associate verbal with visual sequences. (This report to appear in J. Barnden and J. Pollack (Eds.) Advances in Connectionist and Neural Computation Theory. Ablex Publ. An initial version of this report was presented at the AAAI & ONR sponsored Workshop on HIgh-Level Connectionism, held at New Mexico State University, April 9-11, 1988.) For copies of this tech. rep., please send requests to: Valerie@CS.UCLA.EDU or Valerie Aylett 3532 Boelter Hall Computer Science Dept. UCLA, Los Angeles, CA 90024 ------------------------------ End of Neurons Digest *********************