File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/91/p91-1010_metho.xml

Size: 20,133 bytes

Last Modified: 2025-10-06 14:12:48

<?xml version="1.0" standalone="yes"?>
<Paper uid="P91-1010">
  <Title>TYPE-RAISING AND DIRECTIONALITY IN COMBINATORY GRAMMAR*</Title>
  <Section position="4" start_page="0" end_page="72" type="metho">
    <SectionTitle>
(6) Forward Composition:
X/Y Y/Z ~B X/Z (&gt;B)
</SectionTitle>
    <Paragraph position="0"> The rule corresponds to Curry's eombinator B, as the subscripted arrow indicates. It allows sentences like Mary admires, and may enjoy, musicals to be accepted, via the functional composition of two verbs (indexed as &gt;B), to yield a composite of the same category as a transitive verb. Crucially, composition also yields the appropriate interpretation for the composite verb may prefer in this sentence (the rest of the derivation is as in (3)):  (7) admires and may enjoy</Paragraph>
    <Paragraph position="2"> CCG also allows type-raising rules, related to the combinator T, which turn arguments into functions over functions-over-such-arguments. These rules allow arguments to compose, and thereby lake part in coordinations like I dislike, and Mary enjoys, musicals. They too have an invariant compositional semantics which ensures that the result has an appropriate interpretation. For example, the following rule allows such conjuncts to form as below (again, the remainder of the derivation is omitted):</Paragraph>
    <Paragraph position="4"> This apparatus has been applied to a wide variety of phenomena of long-range dependency and coordinate structure (cf. \[2\], \[5\], \[6\]). 1 For example, Dowty proposed to account for the notorious &amp;quot;non-constituent&amp;quot; coordination in (10) by adding two rules that are simply the backward mitre-image versions of the composition and type raising rules already given (they are indicated in the derivation by &lt;B and &lt;T). 2 This is a welcome result: not only do we capture a construction that has been resistant to other formalisms. We also satisfy a prediction of the theory, for the two backward rules arc clearly expected once we have chosen to introduce their mirror image originals. The earlier papers show that, provided type raising is limited to the two &amp;quot;order preserving&amp;quot; varieties exemplified in these examples, the above reduction is the only one permitted by the lexicon of English. A number of related cross-linguistic regularities in the dependency of gapping upon basic word order follow (\[2\], \[6\]).</Paragraph>
    <Paragraph position="5"> The construction also strongly suggests that all NPs (etc.) should be considered as type raised, preferably I One further class of rules, corresponding to the combinator S, has been proposed. This combinator is not discussed here, but all the present results transfer to tho6e rules as well.</Paragraph>
    <Paragraph position="6"> 2This and other long examples have been &amp;quot;flmted&amp;quot; to later positions in the text.</Paragraph>
    <Paragraph position="7"> in the lexicon, and that categories like NP should not reduce at all. However, this last proposal seems tc implies a puzzling extra ambiguity in the lexicon, and for the moment we will continue to view type-raising as a syntactic rule.</Paragraph>
    <Paragraph position="8"> The universal claim depends upon type-raising being limited to the following schemata, which do not of themselves induce new constituent orders:</Paragraph>
    <Paragraph position="10"> If the following patterns (which allow constituent orders that are not otherwise permitted) were allowed, the regularity would be unexplained, and without further restrictions, grammars would collapse into free order:</Paragraph>
    <Paragraph position="12"> But what are the principles that limit combinatory rules of grammar, to include (11) and exclude (12)? The earlier papers claim that all CCG rules must conform to three principles. The first is called the Principle of Adjacency \[5, pA05\], and says that rules may only apply to string-adjacent non-empty categories. It amounts to the assumption that combinatops will do the job. The second is called the Principle of Directional Consistency. Informally stated, it says that rules may not override the directionality on the &amp;quot;cancelling&amp;quot; Y category in the combination. For example, the following rule is excluded:</Paragraph>
    <Paragraph position="14"> The third is the Principle of Directional Inheritance, which says that the directionality of any argument in the result of a combinatory rule must be the same as the directionality on the corresponding argument(s) in the original functions. For example, the following composition rule is excluded:</Paragraph>
    <Paragraph position="16"> However, rules like the following are permitted:</Paragraph>
    <Paragraph position="18"> This rule (which is not a theorem in the Lambek calculus) is used in \[5\] to account for examples like I shall buy today and read tomorrow, the collected works of Proust, the crucial combination being the following:  to the simple statement that combinatory rules may not contradict the directionality specified in the lexicon. But how is this observation to be formalised, and how does it bear on the type-raising rules? The next section answers these questions by proposing an interpretation, grounded in string positions, for the symbols / and \ in CCG. The notation will temporarily become rather heavy going, so it should be clearly understood that this is not a proposal for a new CCG notation. It is a semantics for the metagrammar of the old CCG notation.</Paragraph>
  </Section>
  <Section position="5" start_page="72" end_page="73" type="metho">
    <SectionTitle>
DIRECTIONALITY IN CCG
</SectionTitle>
    <Paragraph position="0"> The fact that directionality of arguments is inherited under combinatory rules, under the third of the principles, strongly suggests that it is a property of arguments themselves, just like their eategorial type, NP or whatever, as in the work of Zeevat et al.</Paragraph>
    <Paragraph position="1"> \[8\]\[9\]. However, the feature in question will here be grounded in a different representation, with significantly different consequences, as follows. The basic form of a combinatory rule under the principle of adjacency is a fl ~ ~,. However, this notation leaves the linear order of ot and fl implicit. We therefore temporarily expand the notation, replacing categories like NP by 4-tuples, of the form {e~, DPa, L~, Ra}, comprising: a) a type such as NP; b) a Distinguished Position, which we will come to in a minute; c) a Leftend position; and d) a Right-end position. The Principle of Adjacency finds expression in the fact that all legal combinatory rules must have the the form in (17), in which the right-end of ~ is the same as the left-end of r: We will call the position P2, to which the two categories are adjacent, the juncture.</Paragraph>
    <Paragraph position="2"> The Distinguished Position of a category is simply the one of its two ends that coincides with the juncture when it is the &amp;quot;'cancelling&amp;quot; term Y. A rightward combining function, such as the transitive verb enjoy, specifies the distinguished position of its argument (here underlined for salience) as being that argument's left-end. So this category is written in full as in (18)a, using a non-directional slash/. The notation in (a) is rather overwhelming. When positional features are of no immediate relevance in such categories, they will be suppressed. For example, when we are thinking of such a function as a function, rather than as an argument, we will write it as in (18)b, where VP stands for {VP, DFVp, Lw,, Rvp}, and the distinguished position of the verb is omitted. It is important to note that while the binding of the NP argument's Distinguished Position to its left hand end L,p means that enjoy is a rightward function, the distinguished position is not bound to the right hand end of the verb, t~verb. It follows that the verb can potentially combine with an argument elsewhere, just so long as it is to the right. This property was crucial to the earlier analysis of heavy NP shift. Coupled with the parallel independence in the position of the result from the position of the verb, it is the point at which CCG parts company with the directional Lambek calculus, as we shall see below.</Paragraph>
    <Paragraph position="3"> In the expanded notation the rule of forward application is written as in (19). The fact that the distinguisbed position must be one of the two ends of an argument category, coupled with the requirement of the principle of Adjacency, means that only the two order-preserving instances of functional application shown in (2) can exist, and only consistent categories can unify with those rules.</Paragraph>
    <Paragraph position="4"> A combination under this rule proceeds as follows.</Paragraph>
    <Paragraph position="5"> Consider example (20), the VP enjoy musicals. The derivation continues as follows. First the positional variables of the categories are bound by the positions in which the words occur in the siring, as in (21), which in the first place we will represent explicitly, as numbered string positions, s Next the combinatory rule (19) applies, to unify the argument term of the function with the real argument, binding the remaining positional variables including the distinguished position, as in (22) and (23). At the point when the combinatory rule applies, the constraint implicit in the distinguished position must actually hold. That is, the distinguished position must be adjacent to the functor.</Paragraph>
    <Paragraph position="6"> Thus the Consistency property of combinatory rules follows from the principle of Adjacency, embodied in the fact that all such rules identify the distinguished position of the argument terms with the juncture P2, the point to which the two combinands are adjacent, as in the application example (19).</Paragraph>
    <Paragraph position="7"> The principle of Inheritance also follows directly from these assumptions. The fact that rules correspond to combinators like composition forces directionality to be inherited, like any other property of an argument such as being an NP. It follows that only instances of the two very general rules of composition shown in (24) are allowed, as a consequence of the three Principles. To conform to the principle of consistency, it is necessary that L~ and /~, the ends of the cancelling category Y, be distinct positions - that is, that Y not be coerced to the empty string. This condition is implicit in the Principle of Adjacency (see above), although in the notation of 3 Declaritivising position like this may seem laborious, but it is a tactic familiar from the DCG literature, from which we shall later borrow the elegant device of encoding such positions implicitly in difference-lists.</Paragraph>
    <Paragraph position="8">  (17) {a, DPa, Px,P~} {\]~,DP~,P2, Ps} ::~ {7, DP.y,P1,Pa} (18) a. enjoy :-- {{VP, DPvp, Lvp, Rvp}/{NP, L.p, Lnp, R.p}, DPverb, Leerb, R~erb} b. enjoy :-- {VP/{NP, Lnp, L.p, P~p}, Leerb, R~erb} (19) {{X, DP., PI, P3}/{Y, P2, P2, P3}, PI, P2} {Y, P2, P2, P3} :~ {X, DPz, PI, P31 (20) 1 enjoy 2 musicals 3 {VP/{NP, Larg, Larg,Rare},Llun,Rlu.} {NP, DPnp, Lnp,R.p} (21) 1 enjoy 2 musicals 3 {VP/{NP, La,,, La,,, R.r,}, 1, 2} {NP, DPnp, 2, 3} (22) I enjoy 2 musicals 3 {VP/{NP, L.rg,Larg,Ro~g},l,2} {NP, DP.p,2,3} {X/{Y, P2, P2, P3}, P1, P2} {Y, P2, P2, P3} (23) 1 enjoy 2 musicals 3 {VP/{NP, 2,2,3~,l,2~ {NP,2,2, 3} {vP, 1, 3} (24) a. {{X, DP~,L.,R.}/{Y, P2,P2,P~},P1,P2} {{Y, P2,P2,P~}/{Z, DPz,Lz,R.},P2,P3) :~ {{X, DPx,L,,,R~,}/{Z, DP.,L.,R.},PI,P3} b. {{Y, P2, Ly, P2}/{Z, DPz, Lz, Rz}, PI, P2} {{X, DPx, L~, R~}/{Y, P2, Lu, P2}, P2, P3} :~ {{X, DPx, Lx,Rz}/{Z, DPz,L,,Rz},PI,P3} (25) The Possible Composition Rules: a. X/Y Y/Z =~B X/Z (&gt;B) b. X/Y Y\Z =~B X\Z (&gt;Bx) e. Y\Z X\Y =~B X\Z (&lt;B) d. Y/Z X\Y ::*'B X/Z (&lt;Bx)  the appendix it has to be explicitly imposed. These schemata permit only the four instances of the rules of composition proposed in \[5\] \[6\], given in (25) in the basic CCG notation. &amp;quot;Crossed&amp;quot; rules like (15) are still allowed Coecause of the non-identity noted in the discussion of (18) between the distinguished position of arguments of functions and the position of the function itself). They are distinguished from the corresponding non-crossing rules by further specifying DP~, the distinguished position on Z. However, no rule violating the Principle of Inheritance, like (14), is allowed: such a rule would require a different distinguished position on the two Zs, and would therefore not be functional composition at all. This is a desirable result: the example (16) and the earlier papers show that the non-order-preserving instances (b, d) are required for the grammar of English and Dutch.</Paragraph>
    <Paragraph position="9"> In configurational languages like English they must of course be carefully restricted as to the categories that may unify with Y.</Paragraph>
    <Paragraph position="10"> The implications of the present formalism for the type-raising rules are less obvious. Type raising rules are unary, and probably lexical, so the principle of adjacency does not apply. However, we noted earlier that we only want the order-preserving instances (11), in which the directionality of the raised category is the reverse of that of its argument. But how can this reversal be anything but an arbitrary property? Because the directionality constraints are grounded out in string positions, the distinguished position of the subject argument of a predicate walks - that is, the right-hand edge of that subject - is equivalent to the distinguished position of the predicate that constitutes the argument of an order-preserving raised sub-ject Gilbert that is, the left-hand edge of that predicate. It follows that both of the order-preserving rules are instances of the single rule (26) in the extended notation: The crucial property of this rule, which forces its instances to be order-preserving, is that the distinguished position variable D Parg on the argument of the predicate in the raised category is the same as that on the argument of the raised category itself. (l'he two distinguished positions are underlined in (26)). Of course, the position is unspecified at the time of applying the rule, and is simply represented as an unbound unification variable with an arbitrary mnemonic identifier. However, when the category combines with a predicate, this variable will be bound by the directionality specified in the predicate itself. Since this condition will be transmitted to the raised category, it will have to coincide with the juncture of the combination. Combination of the categories in the non-grammatical order will therefore fail, just as if the original categories were combining without the mediation of type-raising.</Paragraph>
    <Paragraph position="11"> Consider the following example. Under the above rule, the categories of the words in the sentence Gilbert walks are as shown in (27), before binding.</Paragraph>
    <Paragraph position="12"> Binding of string positional variables yields the categories in (28). The combinatory rule of forward application (19) applies as in example (29), binding further variables by unification. In particular, DP 9, Prop, DPw, and P2, are all bound to the juncture position 2, as in (30). By contrast, the same categories in the opposite linear order fail to unify with any combinatory rule. In particular, the backward application rule fails, as in (31). (Combination is blocked because 2 cannot unify with 3).</Paragraph>
    <Paragraph position="13"> On the assumption implicit in (26), the only permitted instances of type raising are the two rules given earlier as (11). The earlier results concerning word-order universals under coordination are therefore captured. Moreover, we can now think of these two rules as a single underspecified order-preserving rule directly corresponding to (26), which we might write less long-windediy as follows, augmenting the origi- null nal simplest notation with a non-directional slash: (33) The Order-preserving Type-raising Rule:</Paragraph>
    <Paragraph position="15"> The category that results from this rule can combine in either direction, but will always preserve order. Such a property is extremely desirable in a language like English, whose verb requires some arguments to the right, and some to the left, but whose NPs do not bear case. The general raised category can combine in both directions, but will still preserve word order. It thus eliminates what was earlier noted as a worrying extra degree of categorial ambiguity. The way is now clear to incorporate type raising directly into the lexicon, substituting categories of the form T I(TIX), where X is a category like NP or PP, directly into the lexicon in place of the basic categories, or (more readably, but less efficiently), to keep the basic categories and the rule (33), and exclude the base categories from all combination.</Paragraph>
    <Paragraph position="16"> The related proposal of Zeevat et al. \[8\],\[9\] also has the property of allowing a single lexical raised category for the English NP. However, because of the way in which the directional constraints are here grounded in relative string position, rather than being primitive to the system, the present proposal avoids certain difficulties in the earlier treatment. Zeevat's type-raised categories are actually order-changing, and require the lexical category for the English predicate to be S/NP instead of S\NP. (Cf. \[9, pp.</Paragraph>
    <Paragraph position="17">  207-210\]). They are thereby prevented from capturing a number of generalisations of CCGs, and in fact exclude functional composition entirely.</Paragraph>
    <Paragraph position="18"> It is important to be clear that, while the order preserving constraint is very simply imposed, it is nevertheless an additional stipulation, imposed by the form of the type raising rule (26). We could have used a unique variable, DPpr,a say, in the crucial position in (26), unrelated to the positional condition DP~r9 on the argument of the predicate itself, to define the distinguished position of the predicate argument of the raised category, as in example (32).</Paragraph>
    <Paragraph position="19"> However, this tactic would yield a completely unconstrained type raising rule, whose result category could not merely be substituted throughout the lexicon for ground categories like NP without grammatical collapse. (Such categories immediately induce totally free word-order, for example permitting (31) on the English lexicon). It seems likely that type raising is universally confined to the order-preserving kind, and that the sources of so-called free word order lie elsewhere. Such a constraint can therefore be understood in terms of the present proposal simply as a requirement for the lexicon itself to be consistent. It should also be observed that a uniformly order-changing category of the kind proposed by Zeevat et al. is not possible under this theory.</Paragraph>
    <Paragraph position="20"> The above argument translates directly into unification-based frameworks such as PATR or Prolog. A small Prolog program, shown in an appendix, can be used to exemplify and check the argument. 4 The program makes no claim to practicality or efficiency as a CCG parser, a question on which the reader is refered to \[7\]. Purely for explanatory simplicity, it uses type raising as a syntactic rule, rather than as an offline lexical rule. While a few English lexical categories and an English sentence are given by way of illustration, the very general combinatory rules that are included will of course require further constraints if they are not to overgenerate with larger fragments. (For example, &gt;B and &gt;Bx must be disanguished as outlined above, and file latter must be greatly constrained for English.) One very general constraint, excluding all combinations with or into NP, is included in the program, in order to force type-raising and exemplify the way in which further constrained rule-instances may be specified.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML