LAWS@SRI-AI.ARPA (06/24/85)
From: AIList Moderator Kenneth Laws <AIList-REQUEST@SRI-AI> AIList Digest Monday, 24 Jun 1985 Volume 3 : Issue 82 Today's Topics: Queries - VAX Lisp & PC Lisps & McDonnell Douglas NL Breakthrough, Games - Optimal Scrabble, Automata - Predation/Cooperation, Psychology - Common Sense, Analogy - Bibliography, Seminar - Evaluating Expert Forecasts (NASA) ---------------------------------------------------------------------- Date: Mon, 24 Jun 85 07:38:35 EDT From: cugini@NBS-VMS Subject: VAX Lisp Just looking for a little consumer information here - does anyone have any experience with Digital's VAX LISP ? DEC advertises it as a full-fledged implementation of CommonLisp. Any remarks on price, performance, quality, etc are appreciated. John Cugini <Cugini@NBS-VMS> Institute for Computer Sciences and Technology National Bureau of Standards Bldg 225 Room A-265 Gaithersburg, MD 20899 phone: (301) 921-2431 ------------------------------ Date: Sun 23 Jun 85 15:09:12-EDT From: Jonathan Delatizky <DELATZ%MIT-OZ@MIT-MC.ARPA> Subject: PC Lisps [Forwarded from the MIT bboard by SASW@MIT-MC.] Can some of you out there who have used Lisp implementations on IBM PC type machines give me some recommendations as to the best PC Lisp? I plan to run it on a PC/XT and a PC/AT if possible. Also, any expert systems shells that run on the same machines, real or toy-like. ...jon ------------------------------ Date: 22 Jun 1985 13:20-EST From: George Cross <cross%lsu.csnet@csnet-relay.arpa> Subject: McDonnell Douglas NL Breakthrough The following is the text of a full page color ad on page 49 in the June 24, 1985 New Yorker. It has also been run in the Wall Street Journal. Does anyone know what the breakthrough is? This was mentioned on the ailist some time ago but I didn't notice a response. There is a photo of a hand holding the chin of smiling boy. BREAKTHROUGH: A COMPUTER THAT UNDERSTANDS YOU LIKE YOUR MOTHER Having to learn letter-perfect software languages can be frustrating to the average person trying to tap the power of a computer. But practical thinkers at our McDonnell Douglas Computer Systems Company have created the first computer that accepts you as you are - human. They emulated the two halves of the brain with two-level software: One level with a dictionary of facts and a second level to interpret them. The resulting Natural Language processor understands everyday conversational English. So it knows what you mean, no matter how you express yourself. It also learns your idiosyncrasies, forgives your errors, and tells you how to find out what you're looking for. Now, virtually anyone who can read and write can use a computer. We're creating breakthroughs not only in Artificial Intelligence but also in health care, space manufacturing and aircraft. We're McDonnell Douglas. How can I learn more? Write P.O. Box 19501 Irvine, CA 92713 ------------------------------ Date: 22 Jun 1985 13:07-EDT From: Jon.Webb@CMU-CS-IUS2.ARPA Subject: Optimal Scrabble Anyone interested in computer Scrabble should be aware that Guy Jacobson and Andrew Appel (some of the people that did Rog-o-matic) have written a program which in some sense solves the problem. Using a clever data structure, their program makes plays in a few seconds and always makes the best possible play. Their dictionary is the official Scrabble dictionary. The program is not completely optimal because it doesn't take into account how the placement of its words near things like triple word scores may help the other player, but in all other senses it always makes the best play. I suppose some simple strategic techniques could be added using a penalty function, but as the program almost always wins anyway, this hasn't been done. It regularly gets bingos (all seven letters used), makes clever plays that create three or more words, and so on. The version they have now runs on Vax/Unix. There was some work to port it to the (Fat) Macintosh but that is not finished, mainly for lack of interest. Jon ------------------------------ Date: Fri, 21 Jun 85 17:17:58 EDT From: David_West%UMich-MTS.Mailnet@MIT-MULTICS.ARPA Subject: Predation/Cooperation (AIL v3 #78) Re: enquiry of sdmartin@bbng about learning cooperation in predation: For an extensive investigation of a minimal-domain model (prisoner's dilemma),see _The Evolution of Co-operation_ (NY: Basic Books, 1984; LC 83-45255, ISBN 0-465-02122-0) by Robert Axelrod (of the U of Mich). He is in the Institute of Public Policy Studies, but one of his more interesting methods was the use of the genetic algorithms of John Holland (also of the U of Mich) to breed automata to have improved strategies for playing Prisoner's dilemma. A one-sentence summary of his results is that cooperation can displace non-cooperation if individuals remember each other's behavior and have a high enough probability of meeting again. An intermediate-length summary can be found in Science _211_ (27 Mar 81) 1390-1396. ------------------------------ Date: Fri 21 Jun 85 19:23:03-PDT From: Calton Pu <CALTON@WASHINGTON.ARPA> Subject: definition of common sense I had a discussion with a friend on this exact topic just a few weeks ago. My conclusions can be phrased as an elaboration of V. Pratt's two criteria. 1. common knowledge basis (all facts depended on must be common knowledge) I think the (abstract) common knowledge basis can be more concretely described as "cultural background". Your Formico's Pizza example shows clearly that anybody not familiar with San Francisco will not have the "common sense" to go there. The term "cultural background" admits many levels of interpretation (national, provincial, etc.) so most of REALLY COMMON knowledge basis will be encompassed. 2. low computational complexity (easy to check the conclusion). I think the key here is not the checking (NP), but the finding (P) of the solution. So here I differ from Vaughan, in that I believe common sense is something "obvious" to a lot of people, by their own reasoning power. There are two factors involved: the first is the amount of reasoning power; the second is the amount of deductive processing involved. On the first factor, unfortunately usual words to describe people with the adequate reasoning power such as "sensible", "reasonable", and "objective" have also the connotation of being "emotionless". Let's leave out the emotional aspects and use the term "reasonable" to include everybody who is able to apply elementary logic to normal situations. On the second factor, typical words to picture easy deductive efforts are "obvious", "clear", and "evident". So my definition of common sense is: that which is obvious to a reasonable person with an adequate cultural background. I should point out that the three parameters of common sense, cultural background, reasoning power, and deductive effort, vary from place to place and from person to person. If we agreed more on each other's common sense, it might be easier to negotiate peace. ------------------------------ Date: Monday, 24 Jun 85 01:38:08 EDT From: shrager (jeff shrager) @ cmu-psy-a Subject: Analogy Bibliography [Someone asked for an analogy bibliography a while back. This was compiled about two years (maybe more) ago so it's partial and somewhat out of date, but might serve as a starter for people interested in the topic. I've added a couple of thing just now in looking it over. The focus is primarily psychological, but readers will recognize some of the principle AI work as well. I've got annotations for quite a few of these, but the remarks are quite long and detailed so I won't burden AIList with them. -- Jeff] ANALOGY (A partial bibliography) Compiled by Jeff Shrager CMU Psychology 24 June 1985 (Send recommendations to Shrager@CMU-PSY-A.) Bobrow, D. G. & Winograd, T. (1977). An Overview of KRL: A Knowledge Representation Language. Cognitive Science, 1, 3-46. Bott, R.A. A study of complex learning: Theories and Methodologies. Univ. of Calif. at San Diego, Center for Human Information Processing report No. 7901. Brown, D. (1977). Use of Analogy to Acheive New Experience. Technical Report 403, MIT AI Laboratory. Burstein, M. H. (June, 1983). Concept Formation by Incremental Analogical Reasoning and Debugging. Proceedings of the International Machine Learning Workshop. pp. 19-25. Carbonell, J. G. (August, 1981). A computational model of analogical problem solving. Proceedings of the Seventh International Joint Conference on Artificial Intelligence, Vancouver. pp. 147-152. Carbonell, J.G. (1983). Learning by Analogy: Formulating and Generalizing Plans from Past Experience. In Michalski, R.S., Carbonell, J.G., & Mitchell, T.M. (Ed.), Machine Learning, an Aritificial Intelligence Approach Palo Alto: Tioga Press. Carnap, R. (1963). Variety, analogy and periodicity in inductive logic. Philosophy of Science, 30, 222-227. Darden, L. (June, 1983). Reasoning by Analogy in Scientific Theory Construction. Proceedings of the International Machine Learning Workshop. pp. 32-40. de Kleer, J. & Brown, J.S. Foundations of Envisioning. Xerox PARC report. Douglas, S. A., & Moran, T. P. (August, 1983). Learning operator semantics by analogy. Proceedings of the National Conference on Artificial Intelligence. Douglas, S. A., & Moran, T. P. (December, 1983b). Learning text editor semantics by analogy. Proceedings of the Second Annual Conference on Computer Human Interaction. pp. 207-211. Dunker, K. (1945). On Problem Solving. Psychological Monographs, 58, . Evans, T. G. (1968). A program for the solution of a class of geometric analogy intelligence test questions. In Minsky, M. (Ed.), Semantic Information Processing Cambridge, Mass.: MIT Press. pp. 271-253. Gentner, D. (July, 1980). The Structure of Analogical Models in Science. Report 4451, Bolt Beraneck and Newman. Gentner, D. (1981). Generative Analogies as Mental Models. Proceedings of the 3rd National Cognitive Science Conference. pp. 97-100. Proceedings of the 3rd annual conference. Gentner, D. (1982). Are Scientific Analogies Metaphors? In D. S. Miall (Ed.), Metaphor: Problems and Perspectives New York: Harvester Press Ltd. pp. 106-132. Gentner, D., & Gentner, D. R. (1983). Flowing Waters or Teeming Crowds: Mental Models of Electricity. In Gentner, D. & Stevens, A. L. (Ed.), Mental Models Hillsdale, NJ: Lawrence Earlbaum Associates. pp. 99-129. Gick, M. L. & Holyoak, K. J. (1980). Analogic Problem Solving. Cognitive Psychology, 12, 306-355. Gick, M. L. & Holyoak, K. J. (1983). Schema Induction and Analogic Transfer. Cognitive Psychology, 15, 1-38. Halasz, F. & Moran, T. P. (1982). Analogy Considered Harmful. Proceedings of the Conference on Human Factors in Computer Systems, New York. Hesse, Mary. (1955). Science and the Human Imagination. New York: Philisophical Library. Hesse, Mary. (1974). The Structure of Scientific Inference. Berkeley: Univ. of Calif. Press. Kling, R. E. (1971). A Paradigm for Reasoning by Analogy. Artificial Intelligence, 2, 147-178. Lenat, D.B. & Greiner, R.D. (1980). RLL: A representation language language. Proc. of the first annual meeting. Stanford. McDermott, J. (December, 1978). ANA: An assimilating and accomodatiing production system. Technical Report CMU-CS-78-156, Carnegie-Mellon University. McDermott, J. (1979). Learning to use analogies. Sixth Internation Joint Conference on Artificial Intelligence. Medin, D. L. and Schaffer, M. M. (1978). Context Theory of Classification Learning. Psychological Review, 85(3), 207-238. Minsky, M. (1975). A Framework for Representing Knowledge. In Winston, P.H. (Ed.), The Psychology of Computer Vision New York: McGraw Hill. Minsky, M. (July, 1982). Learning Meaning. Technical Report, . Unpublished MIT AI Lab techinical report. Moore, J. & Newell, A. (1974). How can MERLIN Understand? In L.W.Gregg (Ed.), Knowledge and Cognition Potomic, Md.: Erlbaum Associates. Ortony, A. (1979). Beyond Literal Similarity. Psych Review, 86(3), 161-179. Pirolli, P. & Anderson, J.R. (1985) The role of Learning from Examples in the Acquisition of Recursive Programming Skills. Canadian Journal of Psychology. Vol. 39, no. 4; pgs. 240-272. Polya, G. (1945). How to solve it. Princton, N.J.: Princeton U. Press. Quine, W. V. O. (1960). Word and Object. Cambridge: MIT Press. Reed, S. K., Ernst, G. W., & Banerji, R. (1974). The Role of Analogy in Transfer Between Similar Problem States. Cognitive Psychology, 6, 436-450. Rosch, E., Mervis, C. B., Gray, W. D., Johnson, D. M., & Boynes-Braem, P. (1976). Basic Objects on Natural Kinds. Cog Psych, 8, 382-439. Ross, B. (1982). Remindings and Their Effects in Learning a Cognitive Skill. PhD thesis, Stanford. Rumelhart, D.E., & Norman, D.A. (?DATE?). Accretion, tuning, and restructuring: Three modes of learning. In R.Klatsky and J.W.Cotton (Eds.), Semantic Factors in Cognition Hillsdale, N.J.: Erlbaum Associates. Rumerlhart, D.E. & Norman, D.A. (1981). Analogical Processes in Learning. In J.R. Anderson (Ed.), Cognitive Skills and Their Acquisition Hillsdale, N.J.: Lawrence Earlbaum Associates. pp. 335-360. Schustack, M., & Anderson, J. R. (1979). Effects of analogy to prior knowledge on memory for new information. Journal of Verbal Learning and Verbal Behavior, 18, 565-583. Sembugamoorthy, V. (August, 1981). Analogy-based acquisition of utterances relating to temporal aspects. Proceedings of the 7th International Joint Conference on Artificial Intelligence. pp. 106-108. Shrager, J. & Klahr, D. (December, 1983). A Model of Learning in the Instructionless Environment. Proceedings of the Conference on Human Factors in Computing Systems. pp. 226-229. Shrager, J. & Klahr, D. Instructionless Learning: Hypothesis Generation and Experimental Performance. In preparation. Sternberg, R. (1977). Intelligence, information processing, and analogical reasoning: The componential analysis of human abilities. Hillsdale, N.J.: Lawrence Erlbaum Associates. VanLehn, K., & Brown, J. S. (1978). Planning nets: A representation for formalizing analogies and semantic models of procedural skills. In Snow, R. E., Frederico, P. A. and Montague, W. E. (Ed.), Aptitude Learning and Instruction: Cognitive Process Analyses Hillsdale, NJ: Lawrence Erlbaum Associates. Weiner, E. J. A Computational Approach to Metaphore Comprehension. In the Penn Review of Linguistics. Winston, P. H. (December, 1980). Learning and Reasoning by Analogy. Communications of the ACM, 23(12), 689-703. Winston, P. H. Learning and Reasoning by Analogy: The details. MIT AI Memo number 520. ------------------------------ Date: Fri, 21 Jun 85 11:42:26 pdt From: gabor!amyjo@RIACS.ARPA (Amy Jo Bilson) Subject: Seminar - Evaluating Expert Forecasts (NASA) NASA PERCEPTION AND COGNITION SEMINARS Who: Keith Levi From: University of Michigan When: 10 am, Tuesday, June 25, 1985 Where: Room 177, Building 239, NASA Ames Research Center What: Evaluating Expert Forecasts Abstract: Probabilistic forecasts, often generated by an expert, are critical to many decision aids and expert systems. The quality of such inputs has usually been evaluated in terms of logical consistency. However, in terms of real-world implications, the external correspondence of probabilistic forecasts is usually much more important than internal consistency. I will discuss recently developed procedures for evaluating external correspondence and present research on the topic. Non-citizens (except permanent residents) must have prior approval from the Directors Office one week in advance. Permanent residents must show Alien Registration Card at the time of registration. To request approval or obtain further information, call 415-694-6584. ------------------------------ End of AIList Digest ********************