File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/04/c04-1075_metho.xml

Size: 21,475 bytes

Last Modified: 2025-10-06 14:08:40

<?xml version="1.0" standalone="yes"?>
<Paper uid="C04-1075">
  <Title>A High-Performance Coreference Resolution System using a Constraint-based Multi-Agent Strategy</Title>
  <Section position="4" start_page="0" end_page="0" type="metho">
    <SectionTitle>
2 Preprocessing: Determination of
</SectionTitle>
    <Paragraph position="0"/>
    <Section position="1" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
Referring Expressions
</SectionTitle>
      <Paragraph position="0"> The prerequisite for automatic coreference resolution is to obtain possible referring expressions in an input document. In our system, the possible referring expressions are determined by a pipeline of NLP components:  Among them, named entity recognition, part-of-speech tagging and noun phrase chunking apply the same Hidden Markov Model (HMM) based engine with error-driven learning capability (Zhou and Su 2000). The named entity recognition component (Zhou and Su 2002) recognizes various types of MUC-style named entities, that is, organization, location, person, date, time, money and percentage. The HMM-based noun phrase chunking component (Zhou and Su 2000) determines various noun phrases based on the results of named entity recognition and part-of-speech tagging.</Paragraph>
    </Section>
  </Section>
  <Section position="5" start_page="0" end_page="1" type="metho">
    <SectionTitle>
3 Coreference Types
</SectionTitle>
    <Paragraph position="0"> Since coreference is a symmetrical and transitive relation, it leads to a simple partitioning of a set of referring expressions and each partition forms a coreference chain. Although any two referring expressions in the coreference chain is coreferential, some of conference pairs may be direct while others may be indirect since they only become conferential via other referring expressions in the same coreference chain. This indicates that the most recent antecedent of an anaphor in the coreferential chain is sometimes indirectly linked to the anaphor via some other antecedents in the chain. In these indirect cases, we find that the most recent antecedent always contains little information to directly determine the coreference relationship with the anaphor. Generally, direct and informative coreference pairs are much easier to resolve than indirect and less informative ones. In the following example  The italic markables with the same identification symbol are coreferential.</Paragraph>
    <Paragraph position="1"> &amp;quot;Microsoft Corp.&amp;quot;, &amp;quot;its&amp;quot; and &amp;quot;Microsoft&amp;quot; form a coreference chain. Among the three coreference pairs in the chain, 1) The coreference pair between &amp;quot;Microsoft Corp.&amp;quot; and &amp;quot;Microsoft&amp;quot; is direct.</Paragraph>
    <Paragraph position="2"> 2) The coreference pair between &amp;quot;Microsoft Corp.&amp;quot; and &amp;quot;its&amp;quot; is direct.</Paragraph>
    <Paragraph position="3"> 3) The coreference pair between &amp;quot;its&amp;quot; and &amp;quot;Microsoft&amp;quot; is indirect. This coreference pair only becomes coreferential via another referring expression &amp;quot;Microsoft Corp.&amp;quot; Direct resolution of this coreference pair is error-prone and not necessary since it can be indirectly linked by the other two coreference pairs in the coreference chain.</Paragraph>
    <Paragraph position="4"> Therefore, for a given anaphor, we can always safely filter out these less informative antecedent candidates. In this way, rather than finding the most recent antecedent for an anaphor, our system tries to find the most direct and informative antecedent. This also suggests that we can classify coreference types according to the types of anaphors and restrict the possible types of antecedent candidates for a given anaphor type as follows: * Name alias coreference This is the most widespread type of coreference which is realised by the name alias phenomenon.</Paragraph>
    <Paragraph position="5"> The success of name alias coreference resolution is largely conditional on success at determining when one referring expression is a name alias of another referring expression. Here, the direct antecedent candidate of a named entity anaphor can only be the type of named entity. For example, Microsoft Corp. (i) announced its new CEO yesterday. Microsoft (i) said ...</Paragraph>
    <Paragraph position="6"> * Apposition coreference This is the easiest type of coreference. A typical use of an appositional noun phrase is to provide an alternative description for a named entity. For  example Julius Caesar (i), the well-known emperor (i), was born in 100 BC.</Paragraph>
    <Paragraph position="7"> * Predicate nominal coreference Predicate nominal is typically coreferential with the subject. For example, George W. Bush (i) is the president of the United States (i).</Paragraph>
    <Paragraph position="8"> * Pronominal coreference  This is the second widespread type of coreference which is realised by pronouns. Pronominal coreference has been widely studied in literature of traditional anaphora resolution. The direct antecedent candidate of a pronoun anaphor can be any type of referring expressions. For example, Computational linguistics (i) from different countries attended the tutorial. They (i) took extensive note.</Paragraph>
    <Paragraph position="9"> * Definite noun phrase coreference This is the third widespread type of coreference which is realised by definite noun phrases. It has also been widely studied in the literature of traditional anaphora resolution. A typical case of definite noun phrase coreference is when the antecedent is referred by a definite noun phrase anaphor representing either same concept (repetition) or semantically close concept (e.g. synonyms, super-ordinates). The direct antecedent candidate of a definite noun phrase anaphor can be any type of referring expressions except pronouns. For example, Computational linguistics (i) from different countries attended the tutorial. The participants (i) took extensive note.</Paragraph>
    <Paragraph position="10"> * Demonstrative noun phrase coreference This type of coreference is not widespread. Similar to that of definite noun phrase coreference, the direct antecedent candidate of a demonstrative noun phrase anaphor can be any type of referring expressions except pronouns. For example, Boorda wants to limit the total number of sailors on the arsenal ship (i) to between 50 and 60. Currently, this ship (i) has about 90 sailors.</Paragraph>
    <Paragraph position="11"> * Bare noun phrase coreference The direct antecedent candidate of a bare noun phrase anaphor can be any type of referring expressions except pronouns. For example, The price of aluminium (i) siding has steadily increased, as the market for aluminium (i) reacts to the strike in Chile.</Paragraph>
  </Section>
  <Section position="6" start_page="1" end_page="2" type="metho">
    <SectionTitle>
4 Constraint-based Multi-Agent System
</SectionTitle>
    <Paragraph position="0"/>
    <Section position="1" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
for Coreference Resolution
</SectionTitle>
      <Paragraph position="0"> In accordance with the above differentiation of coreference types according to the anaphor types, a constraint-based multi-agent system is developed.</Paragraph>
    </Section>
    <Section position="2" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
4.1 Common Constraint Agent
</SectionTitle>
      <Paragraph position="0"> For all coreference types described in Section 3, a common constraint agent is applied first using following constraints:</Paragraph>
    </Section>
    <Section position="3" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
Morphological agreements
</SectionTitle>
      <Paragraph position="0"> These constraints require that an anaphor and its antecedent candidate should agree in gender and number. These kinds of morphological agreements has been widely used in the literature of anaphora resolution Semantic consistency This constraint stipulates that the anaphor and its antecedent candidate must be consistent in semantics. For example, the anaphor and its antecedent candidate should contain the same sense or the anaphor contains a sense which is parental to the antecedent candidate. In this paper, WordNet (Miller 1990) is used for semantic consistency check.</Paragraph>
      <Paragraph position="1"> For example, IBM (i) announced its new CEO yesterday.</Paragraph>
      <Paragraph position="2"> The company (i) said ...</Paragraph>
    </Section>
    <Section position="4" start_page="1" end_page="1" type="sub_section">
      <SectionTitle>
4.2 Special Constraint Agents
</SectionTitle>
      <Paragraph position="0"> For each coreference type described in Section 3, a special constraint agent is applied next using some heuristic rules mainly based on the accessibility space, which is learnt from the training data as follows: For a given coreference type and a given valid antecedent type, all the anaphors of the given coreference type are identified first from left to right as they appear in the sentences. For each anaphor, its antecedent is then determined using the principle of proximity. If the most recent antecedent candidate has the given antecedent type, meet the morphological agreements and semantic consistency and is in the same coreference chain as the anaphor, this coreference pair is counted as a correct instance for the given conference type and the given antecedent type.</Paragraph>
      <Paragraph position="1"> Otherwise, it is counted as an error instance. In this way, the precision rates of the coreference type over different valid antecedent types and different accessibility spaces are computed as the percentage of the correct instances among all the correct and error instances. Finally, the accessibility space for a given coreference type and a given antecedent type is decided using a precision rate threshold (e.g. 95%).</Paragraph>
      <Paragraph position="2"> * Agent for name alias coreference A named entity is co-referred with another named entity when the formal is a name alias of the latter. This type of coreference has an accessibility space of the whole document. In this paper, it is tackled by a named entity recognition component, as in Zhou and Su (2002), using the following name alias algorithm in the ascending order of complexity: 1) The simplest case is to recognize full identity of strings. This applies to all types of entity names.</Paragraph>
      <Paragraph position="3"> 2) The next simplest case is to recognize the various forms of location names. Normally, various acronyms are applied, e.g. &amp;quot;NY&amp;quot; vs. &amp;quot;New York&amp;quot; and &amp;quot;N.Y.&amp;quot; vs. &amp;quot;New York&amp;quot;. Sometime, partial mention is also applied, e.g.</Paragraph>
      <Paragraph position="4"> &amp;quot;Washington&amp;quot; vs. &amp;quot;Washington D.C.&amp;quot;. 3) The third case is to recognize the various forms of personal proper names. Thus an article on Microsoft may include &amp;quot;Bill Gates&amp;quot;, &amp;quot;Bill&amp;quot; and &amp;quot;Mr. Gates&amp;quot;. Normally, the full personal name is mentioned first in a document and later mention of the same person is replaced by various short forms such as acronym, the last name and, to a less extent, the first name, of the full person name.</Paragraph>
      <Paragraph position="5"> 4) The most difficult case is to recognize the various forms of organizational names. For various forms of company names, consider a)</Paragraph>
    </Section>
    <Section position="5" start_page="1" end_page="2" type="sub_section">
      <SectionTitle>
&amp;quot;International Business Machines Corp.&amp;quot;,
&amp;quot;International Business Machines&amp;quot; and &amp;quot;IBM&amp;quot;;
</SectionTitle>
      <Paragraph position="0"> b) &amp;quot;Atlantic Richfield Company&amp;quot; and &amp;quot;ARCO&amp;quot;. Normally, various abbreviation forms (e.g. contractions and acronym) and dropping of company suffix are applied. For various forms of other organizational names, consider a) &amp;quot;National University of Singapore&amp;quot;, &amp;quot;National Univ. of Singapore&amp;quot; and &amp;quot;NUS&amp;quot;; b) &amp;quot;Ministry of Education&amp;quot; and &amp;quot;MOE&amp;quot;. Normally, acronyms and abbreviations are applied.</Paragraph>
      <Paragraph position="1"> * Agent for apposition coreference  If the anaphor is in apposition to the antecedent candidate, they are coreferential. The MUC-6 and MUC-7 coreference task definitions are slightly different. In MUC-6, the appositive should be a definite noun phrase while both indefinite and definite noun phrases are acceptable in MUC-7.</Paragraph>
      <Paragraph position="2"> * Agent for predicate nominal coreference If the anaphor is the predicate nominal and the antecedent candidate is the subject, they are coreferential. This agent is still under construction. * Agent for pronominal coreference This agent is applied to the most widely studied coreference: pronominal coreference. 6 heuristic rules are learnt and applied depending on the accessibility space and the types of the antecedent candidates:  1) If the anaphor is a person pronoun and the antecedent candidate is a person named entity, they are coreferential over the whole document.</Paragraph>
      <Paragraph position="3"> 2) If the anaphor is a neuter pronoun and the antecedent candidate is an organization named entity, they are coreferential when they are in the same sentence.</Paragraph>
      <Paragraph position="4"> 3) If the anaphor is a neuter plural pronoun and the antecedent candidate is a plural noun phrase, they are coreferential over the whole document.</Paragraph>
      <Paragraph position="5"> 4) If both the anaphor and the antecedent candidate are third person pronouns, they are coreferential over the whole document.</Paragraph>
      <Paragraph position="6"> 5) If both the anaphor and the antecedent candidate are first or second person pronouns, they are coreferential when they are in the same paragraph.</Paragraph>
      <Paragraph position="7"> 6) If both the anaphor and the antecedent  candidate are neuter pronouns, they are coreferential when they are in the same paragraph or the antecedent candidate is in the previous paragraph of the anaphor.</Paragraph>
      <Paragraph position="8"> * Agent for definite noun phrase coreference The agent for definite noun phrase coreference is mainly based on the accessibility space. This agent is based on the following 3 heuristic rules: 1) The definite noun phrase will be coreferential with a named entity if they are in same paragraph or the entity name is in the previous paragraph of the definite noun phrase.</Paragraph>
      <Paragraph position="9"> 2) The definite noun phrase will be coreferential with a named entity if the head word of the definite noun phrase is only modified by the determiner &amp;quot;the&amp;quot;. That is, the definite noun phrase is of type &amp;quot;the HEADWORD&amp;quot;, e.g. &amp;quot;the company&amp;quot;.</Paragraph>
      <Paragraph position="10">  The agent for demonstrative noun phrase coreference is similar to the agent for definite noun phrase coreference except that the anaphor is a demonstrative noun phrase.</Paragraph>
      <Paragraph position="11"> * Agent for base noun phrase coreference This is the most complicated and confusing coreference in MUC coreference task definitions. Although this type of coreference occupies a large portion, it is hard to find heuristic rules to deal with it. In our system, only one heuristic rule is applied: If the anaphor and the antecedent candidate string-match and include at least two words except the determiner, they are coreferential over the whole document.</Paragraph>
      <Paragraph position="12">  The determiners, e.g. &amp;quot;a&amp;quot;, &amp;quot;an&amp;quot; and &amp;quot;the&amp;quot;, are removed from the strings before comparison. Therefore, &amp;quot;the company&amp;quot; string-matches &amp;quot;a company&amp;quot;.</Paragraph>
    </Section>
    <Section position="6" start_page="2" end_page="2" type="sub_section">
      <SectionTitle>
4.3 Common Preference Agent
</SectionTitle>
      <Paragraph position="0"> For a given anaphor, invalid antecedents are first filtered out using the above common constraint agent and the special constraint agent. Then, the strategy has to choose which of the remaining candidates, if any, is the most likely antecedent candidate. In our strategy, this is done through a common preference agent based on the principle of proximity. That is, our common preference agent takes advantages of the relative locations of the remaining antecedent candidates in the text.</Paragraph>
      <Paragraph position="1"> Among the antecedent candidates:  1) First it looks for those occurring earlier in the current sentence, preferring the one that occurs earliest in the natural left-to-right order.</Paragraph>
      <Paragraph position="2"> 2) If there are no antecedent candidates occurring  earlier in the current sentence, look to those occurring in the immediately preceding sentence of the same paragraph, again preferring the one that occurs earliest in that sentence in left-to-right order.</Paragraph>
      <Paragraph position="3"> 3) If nothing comes up, look back at those occurring in the earlier sentences of the same paragraph, moving back a sentence at a time, but now, within a given sentence preferring the most rightward candidate that occurs later in the sentence.</Paragraph>
      <Paragraph position="4"> 4) Finally, if the scope extends back beyond a paragraph boundary, it looks to those that occur in the sentences of the preceding paragraph, again preferring later to earlier occurrences.</Paragraph>
    </Section>
    <Section position="7" start_page="2" end_page="2" type="sub_section">
      <SectionTitle>
4.4 Multi-Agent Algorithm
</SectionTitle>
      <Paragraph position="0"> The coreference resolution algorithm is implemented based on the previous multi-agents.</Paragraph>
      <Paragraph position="1"> First, all the anaphors are identified from left to right as they appear in the sentences. Then, for a given anaphor, 1) All the referring expressions occurred before the anaphor are identified as antecedent candidates.</Paragraph>
      <Paragraph position="2"> 2) The common constraint agent is applied to filter out the invalid antecedent candidates using various general constraints, such as morphological agreements and semantic consistency constraints.</Paragraph>
      <Paragraph position="3"> 3) The corresponding special constraint agent (if exists) is recalled to first filter out indirect and less informative antecedent candidates and then check the validity of the remaining antecedent candidates by using some heuristic rules. In this way, more invalid antecedent candidates are discarded using various special constraints, such as the accessibility space. 4) The antecedent is chosen from the remaining antecedent candidates, if any, using the common preference agent based on the principle of proximity.</Paragraph>
    </Section>
  </Section>
  <Section position="7" start_page="2" end_page="2" type="metho">
    <SectionTitle>
5 Experimentation
</SectionTitle>
    <Paragraph position="0"> Table 1 shows the performance of our constraint-based multi-agent system on MUC-6 and MUC-7 standard test data using the standard MUC evaluation programs while Table 2 gives the comparisons of our system with others using the same MUC test data and the same MUC evaluation programs. Here, the precision (P) measures the number of correct coreference pairs in the answer file over the total number of coreference pairs in the answer file and the recall (R) measures the number of correct coreference pairs in the answer file over the total number of coreference pairs in the key file while F-measure is the weighted harmonic mean of precision and recall:  Table 1 shows that our system achieves F-measures of 73.9 and 66.5 on MUC-6 and MUC-7 standard test data, respectively. The figures outside the parentheses show the contributions of various agents to the overall recall while the figures inside the parentheses show the frequency distribution of various coreference types in the answer file. It shows that the performance difference between MUC-6 and MUC-7 mainly comes from the significant distribution variation of pronominal coreference. It also shows that there are much room for improvement, especially for the types of pronominal coreference and definite noun pronoun resolution. Table 2 shows that our system achieves significantly better F-measures by 3.1~4.8 percent over the best-reported systems (Ng and Cardie 2002). Most of the contributions come form precision gains. Our system achieves significantly better precision rates by 6.7~10.0 percent over the best-reported systems (Ng and Cardie 2002) while keeping recall rates. One reason behind such high performance is the restriction of indirect and less informative antecedent candidates according to the type of the anaphor. Another reason is differentiation of various types of coreference and the use of multi-agents. In this way, various types of coreference are dealt with effectively by different agents according to their characteristics. The recall difference between our system and the RIPPER system in (Ng and Cardie 2002) maybe come from the predicate nominal coreference, which can be easily resolved using a machine learning algorithm, e.g. (Cohen 1995). Completion of the agent for predicate nominal coreference can easily fill the difference.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML