neuron-request@HPLABS.HP.COM (Neuron-Digest Moderator Peter Marvit) (09/28/88)
Neuron Digest Tuesday, 27 Sep 1988 Volume 4 : Issue 8 Today's Topics: Neural Networks for Temporal Recognition David Touretzky on connectionist vs. symbolic models, knowledge rep. Neural Networks and Expert Systems Stephen Hanson on backpropagation algorithm for neural nets Re: temporal domain in vision Re: temporal domain in vision Call for papers: 6th Scandinavian Conference on Image Analysis Send submissions, questions, mailing list maintenance and requests for back issues to "Neuron-request@hplabs.hp.com" ------------------------------------------------------------ Subject: Neural Networks for Temporal Recognition From: bennett@srcsip.UUCP (Bonnie Bennett) Date: 15 Sep 88 18:19:43 +0000 Problem: Recognizing temporal trends for Expert Systems that require inputs like "X is increasing". Or, any info about Neural Nets and Expert Systems together (again, at last.)) Please send responses directly to me. Thanks Bonnie Bennett (612) 782-7381 [[ See also the second message following. Amonsgt others, Jeff Elman (U.C. at San Diego) has done some very interesting work in looking at how neural nets can do serial processing (cf. "Finding Structure in Time" CRL Tech Report 8801, April 1988). In a talk Stephen Hanson (see later message) gave at Stanford last year, he also talked about "sequential associative memories." -PM]] ------------------------------ Subject: David Touretzky on connectionist vs. symbolic models, knowledge rep. From: pratt@zztop.rutgers.edu (Lorien Y. Pratt) Date: 16 Sep 88 13:54:05 +0000 Fall, 1988 Neural Networks Colloquium Series at Rutgers presents David Touretzky Carnegie-Mellon University Room 705 Hill center, Busch Campus Friday September 23, 1988 at 11:10 am Refreshments served before the talk Abstract My talk will explore the relationship between connectionist models and symbolic models, and ask what sort of things we should expect from a connectionist knowledge representation. In particular I'm interested in certain natural language tasks, like prepositional phrase attachment, which people do rapidly and unconsciously but which involve complicated inferences and a huge amount of world knowledge. - -- - ------------------------------------------------------------------- Lorien Y. Pratt Computer Science Department pratt@paul.rutgers.edu Rutgers University Busch Campus (201) 932-4634 Piscataway, NJ 08854 [[ Editor's note: I include these region talks because, although it's unlikely that most readers will be able to attend, the subjects and speakers are usually of general interest. You can then make a personal effort to contact the speaker for further information. -PM]] ------------------------------ Subject: Neural Networks and Expert Systems From: djlinse@phoenix.Princeton.EDU (Dennis Linse) Date: 16 Sep 88 14:18:06 +0000 In article <8704@srcsip.UUCP> bennett@srcsip.UUCP (Bonnie Bennett) writes: >Problem: Recognizing temporal trends for Expert Systems that require >inputs like "X is increasing". Or, any info about Neural Nets and Expert >Systems together (again, at last.)) Just a note to plug a forthcoming paper related to this very subject. I have in my hand a paper entitled "Integration of Knowledge-Based System and Neural Network Techniques for Robotic Control" by David A. Handelman, Stephen H. Lane, and Jack J. Gelfand to be presented at the IFAC Workshop on Artificial Intelligence in Real-Time Control, in Swansea, UK, next week. It is hot off the presses, and quite interesting. The basic idea is that the expert system controls the robot (a two link arm) and teaches the net. The net takes over until some change in the system requires more learning. (The first author is a soon to be Princeton grad and the second other is a recent grad. All three currently work for David Sarnoff Research Center here in Princeton.) Dennis (djlinse@{phoenix,pucc}.princeton.edu or djlinse@pucc.bitnet) ------------------------------ Subject: Stephen Hanson on backpropagation algorithm for neural nets From: pratt@zztop.rutgers.edu (Lorien Y. Pratt) Date: Tue, 20 Sep 88 18:23:53 +0000 Fall, 1988 Neural Networks Colloquium Series at Rutgers Some comments and variations on back propagation ------------------------------------------------ Stephen Jose Hanson Bellcore Cognitive Science Lab, Princeton University Room 705 Hill center, Busch Campus Friday September 30, 1988 at 11:10 am Refreshments served before the talk Abstract Backpropagation is presently one of the most widely used learning techniques in connectionist modeling. Its popularity, however, is beset with many criticisms and concerns about its use and potential misuse. There are 4 sorts of criticisms that one hears: (1) it is a well known statistical technique (least squares) (2) it is ignorant <about the world in which it is learning--thus design of i/o is critical> (3) it is slow--(local minima, its NP complete) (4) it is ad hoc--hidden units as "fairy dust" I believe these four types of criticisms are based on fundamental misunderstandings about the use and relation of learning methods to the world, the relation of ontogeny to phylogeny, the relation of simple neural models to neuroscience and the nature of "weak" learning theories. I will discuss these issues in the context of some variations on backpropagation. P.S. Don't forget the talk this Friday (the 23rd) by Dave Touretzky - -- - ------------------------------------------------------------------- Lorien Y. Pratt Computer Science Department pratt@paul.rutgers.edu Rutgers University Busch Campus (201) 932-4634 Piscataway, NJ 08854 ------------------------------ Subject: Re: temporal domain in vision From: lag@cseg.uucp (L. Adrian Griffis) Date: Wed, 21 Sep 88 20:50:53 +0000 In article <233@uceng.UC.EDU>, dmocsny@uceng.UC.EDU (daniel mocsny) writes: > In Science News, vol. 134, July 23, 1988, C. Vaughan reports on the > work of B. Richmond of NIMH and L. Optican of the National Eye > Institute on their multiplex filter model for encoding data on > neural spike trains. The article implies that real neurons multiplex > lots of data onto their spike trains, much more than the simple > analog voltage in most neurocomputer models. I have not seen > Richmond and Optican's papers and the Science News article was > sufficiently watered down to be somewhat baffling. Has anyone > seen the details of this work, and might it lead to a method to > significantly increase the processing power of an artificial neural > network? My understanding is that neurons in the eye depart from a number of general rules that neurons seem to follow elsewhere in the nervous system. One such departure is that sections of a neuron can fire independent of other sections. This allows the eye to behave as though is has a great many logical neuron without having to use the the space that the same number of discrete cellular metabolic systems would require. I'm not an expert in this field, but this suggests to me that many of the special tricks that neurons of the eye employ may be attempts to overcome space limitations rather than to make other processing schemes possible. Whether or not this affects the applicability of such tricks to artificial neural networks is another matter. After all, artificial neural networks have space limitations of their own. - -- UseNet: lag@cseg L. Adrian Griffis BITNET: AG27107@UAFSYSB ------------------------------ Subject: Re: temporal domain in vision From: jwl@ernie.Berkeley.EDU (James Wilbur Lewis) Date: Thu, 22 Sep 88 01:42:30 +0000 In article <724@cseg.uucp> lag@cseg.uucp (L. Adrian Griffis) writes: >In article <233@uceng.UC.EDU>, dmocsny@uceng.UC.EDU (daniel mocsny) writes: >> In Science News, vol. 134, July 23, 1988, C. Vaughan reports on the >> work of B. Richmond of NIMH and L. Optican of the National Eye >> Institute on their multiplex filter model for encoding data on >> neural spike trains. The article implies that real neurons multiplex >> lots of data onto their spike trains, much more than the simple >> analog voltage in most neurocomputer models. > >My understanding is that neurons in the eye depart from a number of >general rules that neurons seem to follow elsewhere in the nervous system. I think Richmond and Optican were studying cortical neurons. Retinal neurons encode information mainly by graded potentials, not spike trains....another significant difference between retinal architecture and most of the rest of the CNS. I was somewhat baffled by the Science News article, too. For example, it was noted that the information in the spike trains might be a result of the cable properties of the axons involved, not necessarily encoding any "real" information, but this possibility was dismissed with a few handwaves. Another disturbing loose end was the lack of discussion about how this information might be propogated across synapses. Considering that it generally takes input from several other neurons to trigger a neural firing, and that the integration time necessary would tend to smear out any such fine-tuned temporal information, I don't see how it could be. It's an interesting result, but I think they may have jumped the gun with the conclusion they drew from it. - -- Jim Lewis U.C. Berkeley ------------------------------ Subject: Call for papers: 6th Scandinavian Conference on Image Analysis From: <PARKKINE%FINKUO.BITNET@CUNYVM.CUNY.EDU> Date: Thu, 22 Sep 88 15:06:00 +0100 The 6th Scandinavian Conference on Image Analysis ================================================= June 19 - 22, 1989 Oulu, Finland Second Call for Papers INVITATION TO 6TH SCIA The 6th Scandinavian Conference on Image Analysis (6SCIA) will be arranged by the Pattern Recognition Society of Fin- land from June 19 to June 22, 1989. The conference is spon- sored by the International Association for Pattern Recogni- tion. The conference will be held at the University of Oulu. Oulu is the major industrial city in North Finland, situated not far from the Arctic Circle. The conference site is at the Linnanmaa campus of the University, near downtown Oulu. CONFERENCE COMMITTEE Erkki Oja, Conference Chairman Matti Pietik{inen, Program Chairman Juha R|ning, Local organization Chairman Hannu Hakalahti, Exhibition Chairman Jan-Olof Eklundh, Sweden Stein Grinaker, Norway Teuvo Kohonen, Finland L. F. Pau, Denmark SCIENTIFIC PROGRAM The program will consist of contributed papers, invited talks and special panels. The contributed papers will cov- er: * computer vision * image processing * pattern recognition * perception * parallel algorithms and architectures as well as application areas including * industry * medicine and biology * office automation * remote sensing There will be invited speakers on the following topics: Industrial Machine Vision (Dr. J. Sanz, IBM Almaden Research Center) Vision and Robotics (Prof. Y. Shirai, Osaka University) Knowledge-Based Vision (Prof. L. Davis, University of Maryland) Parallel Architectures (Prof. P. E. Danielsson, Link|ping University) Neural Networks in Vision (to be announced) Image Processing for HDTV (Dr. G. Tonge, Independent Broadcasting Authority). Panels will be organized on the following topics: Visual Inspection in the Electronics Industry (moderator: prof. L. F. Pau); Medical Imaging (moderator: prof. N. Saranummi); Neural Networks and Conventional Architectures (moderator: prof. E. Oja); Image Processing Workstations (moderator: Dr. A. Kortekan- gas). SUBMISSION OF PAPERS Authors are invited to submit four copies of an extended summary of at least 1000 words of each of their papers to: Professor Matti Pietik{inen 6SCIA Program Chairman Dept. of Electrical Engineering University of Oulu SF-90570 OULU, Finland tel +358-81-352765 fax +358-81-561278 telex 32 375 oylin sf net scia@steks.oulu.fi The summary should contain sufficient detail, including a clear description of the salient concepts and novel features of the work. The deadline for submission of summaries is December 1, 1988. Authors will be notified of acceptance by January 31st, 1989 and final camera-ready papers will be re- quired by March 31st, 1989. The length of the final paper must not exceed 8 pages. In- structions for writing the final paper will be sent to the authors. EXHIBITION An exhibition is planned. Companies and institutions in- volved in image analysis and related fields are invited to exhibit their products at demonstration stands, on posters or video. Please indicate your interest to take part by con- tacting the Exhibition Committee: Matti Oikarinen P.O. Box 181 SF-90101 OULU Finland tel. +358-81-346488 telex 32354 vttou sf fax. +358-81-346211 SOCIAL PROGRAM A social program will be arranged, including possibilities to enjoy the location of the conference, the sea and the midnight sun. There are excellent possibilities to make post-conference tours e.g. to Lapland or to the lake dis- trict of Finland. The social program will consist of a get-together party on Monday June 19th, a city reception on Tuesday June 20th, and the conference Banquet on Wednesday June 21st. These are all included in the registration fee. There is an extra fee for accompanying persons. REGISTRATION INFORMATION The registration fee will be 1300 FIM before April 15th, 1989 and 1500 FIM afterwards. The fee for participants cov- ers: entrance to all sessions, panels and exhibition; proceedings; get-together party, city reception, banquet and coffee breaks. The fee is payable by - check made out to 6th SCIA and mailed to the Conference Secretariat; or by - bank transfer draft account or - all major credit cards Registration forms, hotel information and practical travel information are available from the Conference Secretariat. An information package will be sent to authors of accepted papers by January 31st, 1989. Secretariat: Congress Team P.O. Box 227 SF-00131 HELSINKI Finland tel. +358-0-176866 telex 122783 arcon sf fax +358-0-1855245 There will be hotel rooms available for participants, with prices ranging from 135 FIM (90 FIM) to 430 FIM (270 FIM) per night for a single room (double room/person). ------------------------------ End of Neurons Digest *********************