nl-kr-request@CS.ROCHESTER.EDU (NL-KR Moderator Brad Miller) (09/01/88)
NL-KR Digest (8/31/88 18:45:37) Volume 5 Number 12 Today's Topics: English grammar (open/closed classes) Re: Category Theory in AI Re: Chomsky reference Speech recognition with neural nets Model-theoretic semantics Submissions: NL-KR@CS.ROCHESTER.EDU Requests, policy: NL-KR-REQUEST@CS.ROCHESTER.EDU ---------------------------------------------------------------------- Date: Sun, 14 Aug 88 01:34 EDT From: Stephen D. Crocker <crocker@tis-w.arpa> Subject: open versus closed classes of words in English grammar McGuire replied to Nagle's query about open versus closed classes of words in English grammar, viz nouns, verbs, adjectives and adverbs are open and conjunctions, articles, prepositions, etc. are closed. He then comments: > While I'm familiar with this distinction, and think that it may have > been around in linguistics for quite some while (Bernard Bloch maybe?), > I don't remember it being used much. The only references that spring to > mind are some studies in speech production and slips of the tongue done > in the 70s by Anne Cunningham (she's a Brit though I'm not sure of her > last name) and maybe Victoria Fromkin claiming that less errors are > associated with closed class words and that they play some privileged role > in speech_production/syntax/lexical_access/the_archetecture_of_the_mind. I recall in the mid or late 60's reading about a parser built in the UK that relied heavily on the closed classes -- I think the term was "functions words". I believe the parser determined which class the other words were in, noun, verb, etc., solely by the slots created from the function words. To that parser, McGuire's four example sentences would be equivalent to "Foo frobbed fie" "Foo has frobbed fie" "Foo might frob fie" "Foo fums to frob fie" The parser was exceedingly fast, but I don't remember any follow up from this work. If pressed, I can probably find a reference, but I suspect many readers of this digest are more familiar with the work than I. In the speech understanding work of the early 70's, I found it interesting that the functions words played a lesser role than might have been expected because they tended to be unstressed when spoken and hence reduced in duration and clarity. I don't recall whether they played a major role in any of the later systems. It's evident that humans depend on these words and learn new open class words from context created by a combination of the closed class words and known meanings for the open class words elsewhere in the sentence. This suggests that one attribute to look for in truly mature speech understanding systems is reliable "hearing" of function words. I'd be interested if anyone knows the current status of speech understanding in this area. Along somewhat separate lines, Balzer at ISI built a rudimentary parser for English in the early 70's. It was aimed at extracting formal program specs from an English specification. His key example was based heavily on interpeting the closed classes and treating the open classes as variables. ------------------------------ Date: Mon, 15 Aug 88 17:42 EDT From: HILLS%reston.unisys.com@RELAY.CS.NET Subject: Re: English Grammar In AI List V8 #35 John Nagle described a grammar which divided words into four catagories and requested a reference for the list of 'special' words. This may be related to the work of Miller, Newman, and Friedman of Harvard. In 1958 they proposed that words should be divided into two classes which they defined as follows: We will call these two classes the "function words" and the "content words". Function words include those which are traditionally called articles, prepositions, pronouns, conjunctions, and auxillary verbs, plus certain irregular forms. The function words have rather specific syntactic functions which must, by and large, be known individually to the speaker of English. The content words include those which are traditionally called nouns, verbs, and adjectives, plus most of the adverbs. It is relatively easy to add new content words to a language, but the set of function words is much more resistant to inovations. The list of function words is included in the book: 'Elements of Software Science' by Maurice H. Hallstead, Elsevier, 1977. This list contains about 330 words. I suspect that the list of 'special words' sought by Nagle is contained within this list of function words. -- Fred Hills ------------------------------ Date: Tue, 23 Aug 88 11:39 EDT From: GA3662%SIUCVMB.BITNET@CUNYVM.CUNY.EDU Subject: English grammar (open/closed classes) From: ga3662@siucvmb ('Geoff Nathan' to humans) Further to John Nagle's question about the concepts of open and closed classes. An excellent summarry of the set and concept for English can be found in an ancient book: Charles C. Fries. The structure of English. Harcourt Brace. 1952. This is a classic structuralist description of English, with things like 'class A, B, 1, 2' etc. replacing traditional 'noun, verb' etc. labels. The distinction is also discussed in such works as Gleason (Intro. to Descriptive Linguistics, and Linguistics and English Grammar). While it has no formal place in most versions of generative grammar, it does figure in Montague semantics (albeit indirectly). Generally, open class items are not provided with a translation, except something like the meaning of 'horse' is {horse'} (notation changed because of the limitations of this keyboard). On the other hand, closed class members such as 'the', 'is', 'may' etc. will get some translation into intensional logic, complete with lambdas etc. Some would say that the distinction is not useful for synchronic descriptions, since it is merely a predictor of where new words are likely to come from. Further to Mcguire's discussion, closed class members are *sometimes* terminal symbols in a PS grammar (as opposed to being inserted under category symbols like N, V etc.) The references to psycholinguistic investigations about this topic are by Ann Cutler. Some of her work may be found in Linguistic Inquiry in articles dealing with what she calls the 'mental lexicon'. Joan Bybee's book 'Morphology' suggests some semantic reasons why languages might put certain categories in closed classes and others in open classes. -- Geoff Nathan (ga3662@siucvmb) Human Address: Department of Linguistics, Southern Illinois University Carbondale, IL, 62901 ------------------------------ Date: Thu, 18 Aug 88 06:23 EDT From: Jack Campin <jack@cs.glasgow.ac.uk> Subject: Re: Category Theory in AI geddis@atr-la.atr.junet (Donald F. Geddis) wrote: >>dpb@philabs.philips.com (Paul Benjamin) writes: >> Some of us here at Philips Laboratories are using universal >> algebra, and more particularly category theory, to formalize >> concepts in the areas of representation, inference and >> learning. >I'm familiar with those areas of AI, but not with category theory (or >universal algebra, for that matter). Can anyone give a short summary for the >layman of those two mathematical topics? And perhaps a pointer as to how >they might be useful in formalizing certain AI concepts. Thanks! A short summary is tricky without knowing your mathematical background and maybe impossible for a real honest-to-goodness layman. A good book to start with is Herrlich and Strecker's, but if you don't know what a group is, forget it. Arbib and Manes' "Arrows, Structures and Functors" is also OK, but mainly applies it to automata theory (not a booming enterprise these days). Category theory generalizes the notions of "set" and "function", or more generally "mathematical structure" and "mapping that preserves that structure" (where the structures might be, say, n-dimensional Euclidean spaces, and the mappings projections, embeddings and other distance-preserving functions). Its aim is to describe classes of mathematical object (groups, topological spaces, partially ordered sets, ...) by looking at the maps between them, and then to describe relationships between these classes. It captures a lot of otherwise indescribable mathematical notions of "unique" or "natural" objects or maps in a class (the empty set, Descartes' construction of the Euclidean plane as the "product" of two lines, the class of all possible strings in an alphabet, ...). The major application of it to computer science so far is in the semantics of higher-order polymorphic type systems (which can't be described in set theory). David Rydeheard and Rod Burstall have just published a book "Computational Category Theory" that describes categorical constructions algorithmically (in Standard ML) and has a useful bibliography. But a lot of computer science literature that uses category theory does not do so in an essential way; the commutative diagrams are just there to give the authors some mathematical street cred. I can't imagine what category theory has to contribute to knowledge representation (though I can just about imagine it helping to describe neural nets in a more abstract way). Can the philabs people say more about what they're up to? -- ARPA: jack%cs.glasgow.ac.uk@nss.cs.ucl.ac.uk USENET: jack@cs.glasgow.uucp JANET:jack@uk.ac.glasgow.cs useBANGnet: ...mcvax!ukc!cs.glasgow.ac.uk!jack Mail: Jack Campin, Computing Science Dept., Glasgow Univ., 17 Lilybank Gardens, Glasgow G12 8QQ, SCOTLAND work 041 339 8855 x 6045; home 041 556 1878 ------------------------------ Date: Mon, 22 Aug 88 11:35 EDT From: Paul Benjamin <dpb@hen3ry.Philips.Com> Subject: Re: Category Theory in AI In article <1572@crete.cs.glasgow.ac.uk> jack@cs.glasgow.ac.uk (Jack Campin) writes: >I can't imagine what category theory has to contribute to knowledge >representation (though I can just about imagine it helping to describe >neural nets in a more abstract way). Can the philabs people say more >about what they're up to? Well, not really, in a public forum. But Mike Lowry of the Kestrel Institute has pointed out that a representation can be viewed as a category, and a shift of representation as a morphism. The question of whether this insight is very productive is open, but at least it gives us a formal notion of representation, and we've built on this some formal notions of abstraction and learning. We'll let you know if this turns out to be fruitful. Paul Benjamin ------------------------------ Date: Fri, 19 Aug 88 19:13 EDT From: T. William Wells <bill@proxftl.UUCP> Subject: Re: Chomsky reference In article <4044@pdn.UUCP> colin@pdn.UUCP (Colin Kendall) writes: : In article <573@proxftl.UUCP>, bill@proxftl.UUCP (T. William Wells) writes: : > In article <6942@bcsaic.UUCP> rwojcik@bcsaic.UUCP (Rick Wojcik) writes: : > : : > : ... There is no evidence that the well-formedness judgments : > : which people actually make are independent of semantics. : > : > People seem to be able to assign syntactic structure to those : > Lewis Carroll poems. : : Only insofar as the semantics may be guessed at. Let's examine the : famous opening lines from the most famous poem, 'Jabberwocky': Which opens the can of worm labeled: "What is semantics?" As the rest of your posting describes, some knowledge of the world is assumed when assigning syntax. I distinguish the semantics associated with the syntactic function of words (being a noun, verb, adjective, etc.) from the semantics associated with the meanings of the words. (B.T.W. Those categories are much too coarse to describe the actual syntactic categories of words; I'd suggest that, for example, each kind of verb form, distinguished by its objects taken and the basic kind of action being described, is a separate syntactic category.) I think of this kind of semantic information as syntactic. I do believe that we make use of semantics while forming syntactic judgements in order to eliminate the ambiguity that would otherwise result from the exclusion of that information. A legitimate counter-argument supposes that, in order to properly assign a syntactic structure to a sentence, one must elaborate these categories to the extent that each contains only one word (actually concept). In that case, the distinction I make is meaningless. ------------------------------ Date: Tue, 23 Aug 88 14:32 EDT From: Antti Ylikoski <ayl%hutds.hut.fi%FINGATE.BITNET@MITVMA.MIT.EDU> Subject: Speech recognition with neural nets In AIList Digest V8 #63, att!chinet!mcdchg!clyde!watmath!watvlsi!watale!dixit@bloom-beacon.mit.edu (Nibha Dixit) writes: >Subject: Speech rec. using neural nets >Is anyody out there looking at speech recognition using neural >networks? There has been some amount of work done in pattern >recognition for images, but is there anything specific being done >about speech? In the Helsinki University of Technology, in the Department of Technical Physics, the group of Professor Teuvo Kohonen has been studying the usage of neural nets for speech recognition for several years. Professor Kohonen gave a talk on their results in the Finnish AI symposium in this year. They have an experimental system which uses a neural net board in a PC. I cannot remember whether the paper is written in English or in Finnish, but should you wish to get the symposium proceedings, contact Finnish Artificial Intelligence Society (FAIS) c/o Dr Antti Hautamaeki HM & V Research Helsinki, Finland I understand Kohonen's results are comparable to other approaches to speech recognition. --- Andy ------------------------------ Date: Fri, 26 Aug 88 17:43 EDT From: kurt geisel <kgeisel@nfsun.UUCP> Subject: Re: Speech rec. using neural nets Teuvo Kohonen describes success at Helsinki University with a speaker-independent neural system which recognizes phonemes (the box spits out phonemes, not words - you would still need a sophisticated parsing stage) in the article "The 'Neural' Phonetic Typewriter" in the March 1988 issue of the IEEE's _Computer_. +--------------------------------------------------------------------------+ | Kurt Geisel, Intelligent Technology Group, Inc. | | Bix: kgeisel | | ARPA: kgeisel%nfsun@uunet.uu.net US Snail: | | UUCP: uunet!nfsun!kgeisel 65 Lambeth Dr. | | Pittsburgh, PA 15241 | | If a rule fires and no one sees it, did it really fire? | +--------------------------------------------------------------------------+ ------------------------------ Date: Sun, 28 Aug 88 21:05 EDT From: Kai-Fu.Lee@SPEECH2.CS.CMU.EDU Subject: Speech rec. using neural nets In response to Nibha Dixit's question about speech recognition using neural networks, I would recommend the following two articles by Richard Lippmann: An Introduction to Computing with Neural Nets, IEEE ASSP Magazine, Vol. 4, No. 2, April 1987. Neural Nets for Computing, IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), April, 1988. The ICASSP conference proceedings contain quite a few interesting papers on speech recognition with neural networks. Kai-Fu Lee Computer Science Department Carnegie Mellon University Pittsburgh, PA 15213 ------------------------------ Date: Sun, 28 Aug 88 03:06 EDT From: Ching-Yuan Tsai <ching@uhccux.uhcc.hawaii.edu> Subject: Model-theoretic semantics I am interested in knowing whether model-theoretic semantics can be used adequately as a tool to describe the meaning of natural languages. At this moment, my opinion is that it cannot because there is so much in human languages that cannot be formalized. For instance, presupposition and propositional attitudes have long been controversial and problematic; and I believe they are still not solved to any satisfactory extent. In fact, I would say pragmatics is an attempt to compensate for the inadequacies of model-theoretic semantics in dealing with natural languages. The main attraction of model-theoretic semantics is its formalization and rigidness. And this alone makes a lot of people believe that they can use it to provide a theory of truth and meaning for natural languages. My opinion above is vague and maybe biased, but I would like to hear any comments, pros or cons, related to model-theoretic semantics. So far, I could only find three papers sharing my view, and they are listed below: LePore, Ernest. 1983. What model theoretic semantics cannot do? Synthese 54, pp. 167-187. Jardine, Nicholas. 1975. Model theoretic semantics and natural language. In Edward L. Keenan ed. Formal Semantics of Natural Language, pp. 219-240. Cambridge: Cambridge Univ. Press. Potts, Timothy C. 1975. Model theory and linguistics. In Edward L. Keenan ed. Formal Semantics of Natural Language, pp. 241-250. Cambridge: Cambridge Univ. Press. ======================================================================== Bitnet: ching@uhccux.bitnet Internet: ching@uhccux.uhcc.hawaii.edu ---- Ching-yuan Ken Tsai ------------------------------ Date: Sun, 28 Aug 88 09:37 EDT From: Greg Lee <lee@uhccux.uhcc.hawaii.edu> Subject: Re: Model-theoretic semantics From article <2309@uhccux.uhcc.hawaii.edu>, by ching@uhccux.uhcc.hawaii.edu (Ching-Yuan Tsai): > I am interested in knowing whether model-theoretic semantics can >be used adequately as a tool to describe the meaning of natural >languages. The model-theoretic theories that I have seen are more appropriately described as theories of types than as theories of semantics, since they make reference to nothing other than types in the interpretations of expressions. That makes them a species of syntactic theory, I take it, so of course they are not adequate to describe meaning. > At this moment, my opinion is that it cannot because there >is so much in human languages that cannot be formalized. Whatever reservations one might have about model-theoretic "semantics", I don't think *this* is a problem. To the extent that facts of language are known, they can be assigned a representation. To the extent that a theory of the facts is specific, it can be formalized. Whether there is any point to formalization is another matter. If you mean that there is much in human languages that is not known or not understood, of course that's true enough. >... In fact, I would say pragmatics is an >attempt to compensate for the inadequacies of model-theoretic >semantics in dealing with natural languages. Supposing 'semantics' to concern the reference of expressions, and 'pragmatics' to concern this as well as how people use expressions, what you say here is true by definition, because the domain of pragmatics properly includes that of semantics. You've just managed to phrase it in a pejorative way. > The main attraction of model-theoretic semantics is its >formalization and rigidness. ... I don't think model-theoretic semantics is formalized in any but an occasional and incidental way. The form of the statements in such theories has no special significance, so far as I know. What do you mean by "rigidness"? If you mean that the theories are somehow constrained so as to have empirical force, well, I don't think so. Did you ever notice anyone propose a counterexample to Montague Grammar? The general theory, I mean, as opposed to example applications, such as the one in PTQ. (But if the compositionality assumption were taken seriously, any multi-morpheme idiom would serve as a counterexample, as would the existence of subject-verb agreement.) Greg, lee@uhccux.uhcc.hawaii.edu ------------------------------ Date: Mon, 29 Aug 88 11:14 EDT From: Rick Wojcik <rwojcik@bcsaic.UUCP> Subject: Re: Model-theoretic semantics In response to ching@uhccux.UUCP (Ching-Yuan Tsai): The following is a major work that takes a stand against the use of model theoretic semantics: George Lakoff. 1987. Women, Fire, and Dangerous Things. Chicago University Press. -- Rick Wojcik csnet: rwojcik@boeing.com uucp: uw-beaver!ssc-vax!bcsaic!rwojcik address: P.O. Box 24346, MS 7L-64, Seattle, WA 98124-0346 phone: 206-865-3844 ------------------------------ End of NL-KR Digest *******************