File Information
File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/06/w06-1515_metho.xml
Size: 16,835 bytes
Last Modified: 2025-10-06 14:10:41
<?xml version="1.0" standalone="yes"?> <Paper uid="W06-1515"> <Title>Sydney, July 2006. c(c)2006 Association for Computational Linguistics Constraint-based Computational Semantics: A Comparison between LTAG and LRS</Title> <Section position="4" start_page="0" end_page="109" type="metho"> <SectionTitle> 2 LTAG semantics </SectionTitle> <Paragraph position="0"> In (Kallmeyer and Romero, 2005), each elementary tree is linked to a semantic representation (a set of Ty2 formulas and scope constraints). Ty2 formulas (Gallin, 1975) are typed l-terms with individuals and situations as basic types. The scope constraints of the form x [?] y specify subordination relations between Ty2 terms. In other words, x [?] y indicates that y is a component of x.</Paragraph> <Paragraph position="1"> A semantic representation is equipped with a semantic feature structure description. Semantic computation is done on the derivation tree and consists of certain feature value equations between mother and daughter nodes in the derivation tree.</Paragraph> <Paragraph position="2"> (1) John always laughs.</Paragraph> <Paragraph position="3"> As an example, see Fig. 1 showing the derivation tree for (1) with semantic representations and semantic feature structure descriptions as node labels. The additional feature equations in this example are depicted using dotted lines. They arise from top-bottom feature identifications parallel to the unifications performed in FTAG (Vijay-Shanker and Joshi, 1988) and from identifications of global features. They yield 1 = x and 4 = l1. Applying these identities to the semantic representations after having built their union leads to (2). The constraint 3 [?] l1 states that l1 : laugh(x) is a component of 3 .</Paragraph> <Paragraph position="4"> (2) john(x), l2 : always( 3 ), l1 : laugh(x),3 [?] l Note that the feature structure descriptions do not encode the semantic expressions one is interested in. They only encode their contributions to functional applications by restricting the argument slots of certain predicates in the semantic representations: They state which elements are contributed as possible arguments for other semantic expressions and which arguments need to be filled. They thereby simulate lambda abstraction and functional application while assembling the semantic representations. To achieve this, a restricted first order logic is sufficient.</Paragraph> <Paragraph position="5"> Semantic computation is local on the derivation tree: The new feature equations that are added depend only on single edges in the derivation tree. Because of this, even with the extension to semantics, the formalism is still mildly context-sensitive.</Paragraph> </Section> <Section position="5" start_page="109" end_page="110" type="metho"> <SectionTitle> 3 LRS </SectionTitle> <Paragraph position="0"> In LRS the feature logic specifies the entire grammar, including well-formed Ty2 terms as semantic representations, and their mode of composition. Instead of the lambda calculus of traditional Montague Grammar, LRS crucially uses a novel distinction between three aspects of the logical representations of signs (external content, internal content, and parts). LRS constraints establish sub-term relationships between pieces of semantic representations within and across signs, thereby specifying the combinatorial properties of the semantics. The subterm or component-of conditions (symbolized as triangleleft) are imposed by grammar principles. Since these principles are descriptions of object-language expressions, they permit the application of various underspecification techniques of computational semantics, although an LRS grammar does not employ underspecified semantic representations, in contrast to LTAG semantics. null Fig. 2 shows an HPSG description of the syntactic tree and the LRS specifications of (1). The syntactic trees in HPSG correspond to the derived trees of LTAG. Since HPSG does not have derivation trees, the LRS principles refer to derived trees. Each word lexically specifies its contribution to the overall meaning of the sentence (P(ARTS)), the part of its semantics which is outscoped by all signs the word combines with (INC(ONT)), and the overall semantic contribution of its maximal projection (EXC(ONT)). Feature percolation principles identify INC and EXC, respectively, along head projections and collect the elements of the PARTS lists of the daughters at each phrase. The combination of the adjunct with a verbal projection introduces two component-of constraints: The EXC of always must be within the EXC of laughs, and the INC of laughs must be in the scope of always. The semantic argument of laughs (john) is identified by subcategorization (not shown in Fig. 2). A closure condition requires that the semantic representation of an utterance use up all and only the PARTS contributions of all signs, which yields 4 = always(laugh(john)).</Paragraph> </Section> <Section position="6" start_page="110" end_page="111" type="metho"> <SectionTitle> 4 Quantifier scope </SectionTitle> <Paragraph position="0"/> <Section position="1" start_page="110" end_page="110" type="sub_section"> <SectionTitle> 4.1 Specifying a scope window </SectionTitle> <Paragraph position="0"> (3) Exactly one student admires every professor: [?] > [?],[?] > [?] (4) John seems to have visited everybody: seem > [?],[?] > seem Quantificational NPs in English can in principle scope freely (see (3) and (4)). An analysis of quantifier scope must guarantee only two things: 1. the proposition to which a quantifier attaches must be in its nuclear scope, and 2. a quantifier cannot scope higher than the next finite clause. One way to model this is to define a scope window delimited by a maximal scope and a minimal scope for a quantifier. Both LTAG and LRS, specify such scope windows for quantifiers. We will now outline the two analyses.</Paragraph> <Paragraph position="1"> (5) Everybody laughs.</Paragraph> <Paragraph position="2"> (Kallmeyer and Romero, 2005) use global features MAXS and MINS for the limits of the scope window. Fig. 3 shows the LTAG analysis of (5). The feature identifications (indicated by dotted lines) lead to the constraints 2 [?] 5, 5 [?] l1. These constraints specify an upper and a lower boundary for the nuclear scope 5 . With the assignments following from the feature identifications we obtain the semantic representation (6):</Paragraph> <Paragraph position="4"> There is one possible disambiguation consistent with the scope constraints, namely 2 - l2, 4 - l3, 5 - l1. This leads to the semantics every(x, person(x), laugh(x)).</Paragraph> <Paragraph position="5"> In LRS, the EXCONT value of the utterance is the upper boundary while the INCONT value of the syntactic head a quantifier depends on is the lower boundary for scope, as illustrated in Fig. 4. The upper boundary is obtained through the interaction of 1) a PROJECTION PRINCIPLE stating that the</Paragraph> <Paragraph position="7"> each non-head contains its own EXCONT, and b) in an utterance, everything on the PARTS list is a component of the EXCONT. This leads to the constraint 4 a1 6 in Fig. 4, among others. The lower boundary is obtained from the SEMANTICS PRINCIPLE which states that if the non-head of a headed phrase is a quantifier, then the INCONT of the head is a component of its nuclear scope. This yields The striking similarity between the two analyses shows that, despite the fundamental differences between the frameworks, central insights can be modelled in parallel.</Paragraph> </Section> <Section position="2" start_page="110" end_page="111" type="sub_section"> <SectionTitle> 4.2 Nested quantifiers </SectionTitle> <Paragraph position="0"> The use of the upper limit of the scope windows is, however, slightly different: EXCONT contains the quantifier itself as a component while MAXS limits only the nuclear scope, not the quantifier. Consequently, in LTAG the quantifier can scope higher than the MAXS limiting its nuclear scope but in this case it takes immediate scope over the MAXS. (7) Two policemen spy on someone from every city: [?] > [?] > 2 (among others) The LTAG analysis is motivated by nested quantifiers. In sentences such as (7), the embedded quantifier can take scope over the embedding one but if so, this must be immediate scope. In other words, other quantifiers cannot intervene. In (7), the scope order [?] > 2 > [?] is therefore not possible.1 The LTAG analysis is such that the maximal nuclear scope of the embedded quantifier is the propositional label of the embedding quantifier.2 null In LRS, the way the scope window is specified, a corresponding constraint using the EXCONT of the embedded quantifier cannot be obtained.</Paragraph> <Paragraph position="1"> The LRS principle governing the distribution of embedded quantifiers in complex NPs states directly that in this syntactic environment, the embedded quantifier may only take direct scope over the quantifier of the matrix NP. This principle does not refer to the notion of external content at all. At this point it is an open question whether LRS could learn from LTAG here and adapt the scope window so that an analogous treatment of nested quantifiers would be possible.</Paragraph> <Paragraph position="2"> 5 LTAG's extended domain of locality Whereas the treatment of quantification sketched in the preceding section highlights the similarities between LTAG semantics and LRS, this and the following section will illustrate some fundamental differences between the frameworks.</Paragraph> <Paragraph position="3"> In spite of the parallels mentioned above, even INCONT and MINS differ sometimes, namely in sentences containing bridge verbs. This is related to the fact that LTAG has an extended domain of locality whereas HPSG does not. Let us illustrate the difference with the example (8).</Paragraph> <Paragraph position="4"> (8) Mary thinks John will come.</Paragraph> <Paragraph position="5"> 1(Joshi et al., 2003) propose an extra mechanism that groups quantifiers into sets in order to derive these constraints. (Kallmeyer and Romero, 2005) however show that these constraints can be derived even if the upper limit MAXS for nuclear scope is used as sketched above.</Paragraph> <Paragraph position="6"> 2Note that this approach requires constraints of the form l [?] n with l being a label, n a variable. This goes beyond the polynomially solvable normal dominance constraints (Althaus et al., 2003). This extension, though, is probably still polynomially solvable (Alexander Koller, personal communication).</Paragraph> <Paragraph position="7"> In LTAG, the two elementary verb trees (for thinks and will come) have different global MINS features. The one for thinks is the label of the think proposition while the one for will come is the label of the embedded proposition. As a consequence, a quantifier which attaches to the matrix verb cannot scope into the embedded clause. This distinction of different MINS values for different verb trees is natural in LTAG because of the extended domain of locality.</Paragraph> <Paragraph position="8"> In LRS, all verbal nodes in the constituent structure of (8) carry the same INCONT value, namely the proposition of the embedded verb. Consequently, the minimal scope of quantifiers attaching either to the embedding or to the embedded verb is always the proposition of the embedded verb.</Paragraph> <Paragraph position="9"> However, due to the requirement that variables be bound, a quantifier binding an argument of the embedding verb cannot have narrow scope over the embedded proposition.</Paragraph> <Paragraph position="10"> How to implement the LTAG idea of different INCONT values for the embedding and the embedded verb in LRS is not obvious. One might introduce a new principle changing the INCONT value at a bridge verb, whereby the new INCONT would get passed up, and the embedded INCONT would no longer be available. This would be problematic: Take a raising verb as in (9) (adjoining to the VP node in LTAG) instead of a bridge verb: (9) Most people seem to everybody to like the film.</Paragraph> <Paragraph position="11"> Here the minimal scope of most people should be the like proposition while the minimal scope of everybody is the seem proposition. In LTAG this does not pose a problem since, due to the extended domain of locality, most people attaches to the elementary tree of like even though the seem tree is adjoined in between. If the INCONT treatment of LRS were modified as outlined above and seem had an INCONT value that differed from the INCONT value of the embedded like proposition, then the new INCONT value would be passed up and incorrectly provide the minimal scope of most people. LRS must identify the two INCONTs.</Paragraph> <Paragraph position="12"> The difference between the two analyses illustrates the relevance of LTAG's extended domain of locality not only for syntax but also for semantics.</Paragraph> </Section> </Section> <Section position="7" start_page="111" end_page="113" type="metho"> <SectionTitle> 6 Negative Concord </SectionTitle> <Paragraph position="0"> The analysis of negative concord in Polish described in this section highlights the differences in the respective implementation of underspecification techniques in LTAG and LRS. Recall that both LTAG and LRS use component-of constraints. But in LTAG, these constraints link actual Ty2-terms (i.e., objects) to each other, while in LRS, these constraints are part of a description of Ty2-terms.</Paragraph> <Paragraph position="1"> (10) Janek nie pomaga ojcu.</Paragraph> <Paragraph position="2"> Janek NM helps father 'Janek doesn't help his father.' (11) a. Janek nie pomaga nikomu.Janek NM helps nobody 'Janek doesn't help anybody.' b. [?]Janek pomaga nikomu.</Paragraph> <Paragraph position="3"> (12) Nikt nie przyszedl.</Paragraph> <Paragraph position="4"> nobody NM came 'Nobody came.' The basic facts of sentential negation and negative concord in Polish are illustrated in (10)-(12): The verbal prefix nie is obligatory for sentential negation, and it can co-occur with any number of n-words (such as nikt, 'anybody') without ever leading to a double negation reading. As a consequence, (12) expresses only one logical sentential negation, although the negation prefix nie on the verb and the n-word nikt can carry logical negation alone in other contexts. LRS takes advantage of the fact that its specifications of semantic representations are descriptions of logical expressions which can, in principle, mention the same parts of the expressions several times. Fig. 5 shows that both nikt and the verb nie przyszedl introduce descriptions of negations ( 4 and 2 , respectively). The constraints of negative concord in Polish will then conspire to force the negations contributed by the two words to be the same in the overall logical representation 6 of the sentence.</Paragraph> <Paragraph position="5"> Such an analysis is not possible in LTAG. Each negation in the interpretation corresponds to exactly one negated term introduced in the semantic representations. Therefore, the negative particle nie necessarily introduces the negation while the n-word nikt requires a negation in the proposition it attaches to. An analysis along these lines is sketched in Fig. 6 (&quot;GL&quot; stands for &quot;GLOBAL&quot;). The requirement of a negation is checked with a feature NEG indicating the presence of a negation. The scope of the negation (feature N-SCOPE) marks the maximal scope of the existential quantifier of the n-word nikt (constraint 7 [?] 6 ).3 This example illustrates that the two frameworks differ substantially in their treatment of underspecification: 1. LRS employs partial descriptions of fully specified models, whereas LTAG generates underspecified representations in the style of (Bos, 1995) that require the definition of a disambiguation (a &quot;plugging&quot; in the terminology of Bos). 2. LRS constraints contain not Ty2 terms but descriptions of Ty2 terms. Therefore, in contrast to LTAG, two descriptions can denote the same formula. Here, LTAG is more limited compared to LRS. On the other hand, the way semantic representations are defined in LTAG guarantees that they almost correspond to normal dominance constraints, which are known to be polynomially parsable. The difference in the use of underspecification techniques reflects the more general difference between a generative rewriting system such as LTAG, in which the elements of the grammar are objects, and a purely description-based formalism such as HPSG, in which token identities between different components of linguistic structures are natural and frequently employed.</Paragraph> </Section> class="xml-element"></Paper>