nl-kr-request@CS.ROCHESTER.EDU (NL-KR Moderator Brad Miller) (03/01/88)
NL-KR Digest (2/29/88 23:51:29) Volume 4 Number 22 Today's Topics: Re: Garden-path sentences Re: Ambiguity in 1st person plural Are pidgins Indo-European? German-to-English software S. Johnson ==> N. Webster gaffe Neural Network Grammars Re: what is a grammar (for) uses of grammar etc. Submissions: NL-KR@CS.ROCHESTER.EDU Requests, policy: NL-KR-REQUEST@CS.ROCHESTER.EDU ---------------------------------------------------------------------- Date: Thu, 18 Feb 88 18:41 EST From: Martin Taylor <mmt@dciem.UUCP> Subject: Re: Garden-path sentences I *heard* a fine garden path sentence on the BBC news the other day: "In the second test of public opinion ...." (referring to the NH primary) I heard "In the second Test ...", which to one starved for cricket news demands a completion such as " ... England scored 350 for 3 wickets.." (note the upper-case T). The "...of public opinion..." caused a mind-jerking double-take. Note that both versions almost demand that they be heard in the context of a News broadcast, but the second is more likely than the first in an English broadcast, whereas it would be almost impossible in a US broadcast. -- Martin Taylor ....uunet!{mnetor|utzoo}!dciem!mmt mmt@zorac.arpa Magic is just advanced technology ... so is intelligence. Before computers, the ability to do arithmetic was proof of intelligence. What proves intelligence now? Obviously, it is what we can do that computers can't. ------------------------------ Date: Mon, 22 Feb 88 10:08 EST From: Alex Colvin <mac3n@babbage.acc.virginia.edu> Subject: Re: Ambiguity in 1st person plural New Guinea Pidgin (Tok Pisin) has a IE (specifically, English and French) vocabulary, but that doesn't make it an IE language. The grammar isn't IE, but resembles its neighbors, Fijian, Tongan, Samoan, Tahitian, etc, all of which have inclusive/exclusive 1st pl. All but Tok Pisin have dual pronouns. Tok Pisin often sticks the numerals tu, tri into the pronouns, "mitupela" we (excl.) two, "yumitupela", we (incl.) two, "yutripela" (you three), Other non-IE features using IE words are use of -pela (<fellow) as a quantintifier suffixed to a numeral preceding a noun ("sikspela man na meri", six men and women). Taken from "Languages of Asia and the Pacific", Charles Hamblin. Oddly enough, it doesn't include Maaori. I guess he assumes you can get by in English. "Those French, it's like they have a different word for everything!" ------------------------------ Date: Tue, 23 Feb 88 16:14 EST From: Robin Jeffries <jeffries%hplrmj%hplabs@hplabs.HP.COM> Subject: Re: Ambiguity in 1st person plural >Among all the positive anwers to the original question as to whether >there were any languages that distinguish first-person-plural inclusive >from exclusive, no-one has mentioned any Indo-European languages. Are >there any that make this distinction? All the positive answers have been >from around the Indian Ocean or N. American, so far as I remember. This is not a perfect example of this distinction, but I think it at least meets the flavor of what is wanted. Russian makes the distinction at the phrase level. 'myi' (excuse me, but I don't recall how it is supposed to be transliterated) is ambiguously inclusive or exclusive 'we', but tends to mean exclusive, because 'myi c vami' (literally, 'we with you') is used for the inclusive we. 'c vami' is enough of a frozen phrase that I think of these as being two different pronouns that capture exactly the inclusive/exclusive distinction. Robin Jeffries Disclaimer: I am not a linguist, and my Russian is pretty rusty these days, so take all this with a small grain of salt. ------------------------------ Date: Wed, 24 Feb 88 13:01 EST From: WHALEN <jsw@whuts.UUCP> Subject: Re: Ambiguity in 1st person plural In article <2372@csli.STANFORD.EDU>, goldberg@csli.STANFORD.EDU (Jeffrey Goldberg) writes: > In article <2644@dciem.UUCP> mmt@dciem.UUCP (Martin Taylor) writes: > > >Among all the positive anwers to the original question as to whether > >there were any languages that distinguish first-person-plural inclusive > >from exclusive, no-one has mentioned any Indo-European languages.... I was wondering, having just joined the net, whether anyone had mentioned the origin of nosotros in Spanish as opposed to nos and/or nous and nous autres in French. I don't know what the original meaning of nosotros was (? nos altros) but it seems to have implied something different from the original nos. Jon Whalen @ AT&T Bell Labs, Whippany NJ ihnp4!pancho!jsw pancho!jsw@ulysses.att.com ------------------------------ Date: Thu, 25 Feb 88 06:26 EST From: Klaus Schubert <mcvax!dlt1!schubert@uunet.UU.NET> Subject: Are pidgins Indo-European? In the discussion about ambiguity in 1st person plural Erland Sommarskog writes: >> Actually one such a case have already been mentioned, at least if we >> count Tok Pisin to the Indo-Eurpean languages. Tok Pisin is an English- >> based pidign from Papua New Guinea. In Tok Pisin we have: >> "yumi" = "you me" and "mipela" = "me fellow(s)". >> We yet have to see for an "ancient" language of the IE family with >> this distinction though. >> >> Erland Sommarskog >> sommar@enea.UUCP Is there any reason to classify a pidgin language as a member of the Indo- European family one of its source languages belongs to? Isn't this an overestimation of the Indo-European contribution to the new language? Isn't it merely ethnocentrism or laziness when we do not even bother to know what the source languages of a pidgin are called labelling the language as "English-based", "French-based" etc.? Or are there valid linguistic reasons? Klaus Schubert schubert@dlt1.uucp ------------------------------ Date: Tue, 23 Feb 88 15:24 EST From: Debbie Morley <morley@ncrcam.Cambridge.NCR.COM> Subject: German-to-English software Does anyone know of a Unix or MS-DOS software package that does German-English translation? I am interested in a word-to-word dictionary (i.e., enter one German word, receive the English translation), not sentence-to-sentence. Please post, or send any information directly to me. Any information will be appreciated, particularly on how I may acquire such a package. Debbie.Morley@Cambridge.NCR.COM (614)439-0566 ------------------------------ Date: Tue, 23 Feb 88 17:11 EST From: Bruce E. Nevin <bnevin@cch.bbn.com> Subject: S. Johnson ==> N. Webster gaffe I referred to a dictionary informed with the religious and political opinions of Samuel Johnson. Johnson was opinionated, but it was not his Dixionary that I had in mind. I meant, of course, Noah Webster. Bruce Nevin bn@cch.bbn.com <usual_disclaimer> ------------------------------ Date: Thu, 25 Feb 88 00:56 EST From: Ken Laws <LAWS@IU.AI.SRI.COM> Subject: Neural Network Grammars In the last issue it was suggested that thoughts existing prior to verbal expression may be syntactically coded with the aid of a grammar. It was also suggested that natural language may itself be a primary form of thought, in which case syntax and semantics are not independent. I would like to suggest a third viewpoint, one more closely related to spreading-activation networks and perhaps other neural-network representations. Suppose I am "groping for words" during a conversation. It seems unlikely that I am just having difficulty bringing forth pre-existing phrases, but I find it almost as improbable that I have a fully formed thought that is resisting syntactic coding. What I find most plausible is that I have a set of activated concepts that I am trying to linearize in order to produce a sentence. It's rather like trying to string beads. Several orderings (e.g., passive vs. active construction) may be competing, and each implies the activation of additional syntactic elements. This introduces a reinforcement process that usually causes one particular ordering to win, but failure to achieve such dominance can lead to the groping phenomenon. Similar competitions (including goals as well as concepts) govern word choice and the details of articulation. Typing errors offer a particularly good way to study such competitive/cooperative (or relaxation, or constraint-satisfaction) processes. Sometimes I misplace a letter by several character positions. This indicates to me that a neural circuit "knew" that letter had to be included, so it kept trying until it found a place for the insertion. The inappropriateness of the final location was balanced by the increasing urgency felt by this circuit, with a contribution from another circuit trying to enforce the right length for the word. Other factors, such as the ease of making particular letter transitions, may also be active. The final result is thus a composite (or, loosely, a vector sum) of all the constructive and censoring forces that are active during the typing. This model of cognition has some hope of explaining both slips of the tongue and our ease of producing and understanding "ungrammatical" utterances. I doubt, however, that grammars for such processes are useful, feasible, or even meaningful. -- Ken Laws ------------------------------ Date: Thu, 25 Feb 88 11:14 EST From: HESTVIK%BRANDEIS.BITNET@MITVMA.MIT.EDU Subject: Re: what is a grammar (for) MT> Date: Sun, 7 Feb 88 15:29 EST MT> From: Martin Taylor <mmt@dciem.UUCP> MT> Subject: What is a grammar (for) MT> . . . a grammar (a) is a theory MT> about the nature of real language, or (b) [is] a mathematics with its own MT> axioms and procedures for developing theorems, which by chance might MT> parallel some things that are observed in real language. For those working on linguistic theory as a science dealing with an aspect of human psychology, (a) is what is understood as generative grammar. It would be absurd to do science in a way that "by chance" might parallel something in the real world. One qualification is in place: The theory is not about a "real language", it is about a "real grammar". "Language" is partly a product (or reflection) of a grammar. The distiction between the language and the grammar is like the distinction between data and theory in any science. MT> . . . view that MT> a grammar constitutes a set of rules that determine whether a particular MT> string of words is grammatical or not within a language. If her view is MT> correct, then what is a grammar FOR? What would it matter whether a string MT> is grammatical or not? A grammar which only tells us in a yes/no fashion whether something is a wellformed expressions (i.e. the case in the [b]-sense of grammar above) would be of no use in understanding natural language. We want the grammar to tell us *why* a string is grammatical or ungrammatical, i.e. the grammar should give a structural description (an analysis) of both well-formed AND ill-formed expressions. Arild Hestvik Linguistics and Cognitive Science Program Department of Psychology Brandeis University Waltham, MA 02254 ------------------------------ Date: Thu, 25 Feb 88 12:13 EST From: John Nerbonne <nerbonne%hpljan@hplabs.HP.COM> Subject: uses of grammar etc. In Reply to: Date: Tue, 16 Feb 88 18:13 EST From: Bruce E. Nevin <bnevin@cch.bbn.com> Subject: uses of grammar etc (in NL-KR Digest 4 (20), 2/23/88) BN> [...] many formalisms that have developed from the Generative BN> paradigm but do not give themselves the Generative trademark: GPSG, BN> HDPSG, unification grammars, categorial grammars, and so on and on. BN> These formalisms have in common that they overgenerate, that is, they BN> produce structures that do not occur in natural language and must be BN> pruned or filtered out. 1. What generates? A little basic terminology: a grammar formalism specifies the allowable form of a grammar, its rules, whether categories are primitive or defined, etc. A collection of rules etc. in a formalism is a grammar, and it {\it generates} strings of symbols and/or structural analyses of those strings. The goal of (English) descriptive generative linguistics is to find a grammar that generates just the strings and structures we accept as English. But to speak of a formalism as (over)generating is a category error. The formalism doesn't generate at all--over-generate or under-generate or otherwise. It just allows the formulation of rules and grammars whose job is to generate. Furthermore, it is trivial to avoid overgeneration within any formalism; just be very cautious in what you include in the grammar. It's getting things just right that's hard. 2. A Diatribe on "Filtering" But the remarks above are misleading beyond their imprecision. The use of SYNTACTIC FILTERS is controversial in linguistics. A syntactic filter is a device within a grammer that essentially allows an otherwise correct generation to abort. Because filters are suspiciously powerful devices, they ARE NOT USED (outside of MIT's Government and Binding theory). (A final qualification about unification grammars might be worthwhile: unification is a technique for combining feature complexes, and it certainly isn't incompatible with the use of filters. But unification-based theories such as LFG don't use filters.) On the other hand, NO ONE expects syntax to explain every sort of acceptability variation. Some ill-formedness is due to semantics ("a four-legged attitude") and pragmatics ("Don't hear me!"), and some researchers refer to this as a "filtering" of syntax. This is a good idea, and universally accepted. 3. The charge of overgeneration Now, no one has produced a grammar that generates English exactly; they all either overgenerate (generate strings/structures that aren't English) or they undergenerate (fail to generate some acceptable English) or both. It's probably worth mentioning that Chomsky introduced the notion "generative" in {\it Syntactic Structures} as referring to any attempt to describe the syntax of natural language in a way complete enough to define precisely what is in the language. He seems to have assumed that this must be a generating device, or that it might as well be (which assumption has since been discredited, of course, but not too easily). This slant on "generation" clarifies BN's charge: a grammar overgenerates just in case its specifications for English are too inexact. The postulate that noun phrases followed by verb phrases are English sentences is easily seen to "overgenerate" since it allows cases of number discord to count as sentences, etc. Put more polemically: the problem of overgeneration is the same problem as that of achieving the goals of generative theory, that of an exact description of language. 4. An Alternative? BN> The Constructive Grammar of Harris (construction-reduction grammar, BN> composition-reduction grammar, operator-argument grammar) is an example BN> of [...] a mathematical theory of precisely those relations and BN> operations that suffice for language. It does not overgenerate. I can only take this to mean (a) that Con G doesn't attempt a generative characterization at all, in which case the news that it doesn't overgenerate is tautological, and not tidings of comfort or joy; or (b) that it can provide a generative characterization of English syntax. Since this would be a solution (even though not a unique solution) to all the problems of descriptive syntax, my confidence in the scholarly community is such that I think would have heard more of the details by now. I just don't believe (b). Con G must be a different game, so to speak. 5. The purpose of Grammar and "Generative Grammar" A grammar that is generative in Chomsky's sense (a complete characterization of what is in the language) is of value to, say NL Understanding systems, since it allows one to arrive more surely at the genuinely available analyses of NL input. It is a partial solution to the problem of multiple parses in NLU. The problem doesn't thereby go away, but is reduced. Harris's work sounds to me also useful in this regard, and useful roughly in proportion to how successfully it meets generative goals. Generative grammars are less practically useful in other ways: in characterizing a complex human ability, in making traditional grammars more precise. --John Nerbonne nerbonne@hplabs.hp.com ------------------------------ End of NL-KR Digest *******************