File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/93/x93-1014_metho.xml

Size: 15,124 bytes

Last Modified: 2025-10-06 14:13:31

<?xml version="1.0" standalone="yes"?>
<Paper uid="X93-1014">
  <Title>ABCD + E A A andB A andB A,B,C,D A,B,C,D B C</Title>
  <Section position="4" start_page="135" end_page="135" type="metho">
    <SectionTitle>
3. TEMPLATE CORPORA
</SectionTitle>
    <Paragraph position="0"> In order to provide the system developers with training data to illustrate the task and benchmark their develop ment, filled-out templates for the approximately 1000 documents of each training set were provided as &amp;quot;keys&amp;quot;. In addition, templates were produced for the initial TIPSTER program test cycles (12 and 18 months) and for the final joint TIPSTER (24 month)/MUC-5 test. Table 1 provides the number of templates in each development and test set The templates were filled by experienced human analysts according to the same fill rules document (see below) and other supporting documentation that was provided to the system developers to define the exact syntax and semantics of the template fills.</Paragraph>
  </Section>
  <Section position="5" start_page="135" end_page="135" type="metho">
    <SectionTitle>
4. FILL RULES
</SectionTitle>
    <Paragraph position="0"> In addition to the template definition itself, which only defines the syntax in a BNF-like notation (see &amp;quot;Template Design for Information Extraction&amp;quot; in this volume), the analysts and participants in TIPSTER and MUC-5 were provided with fill rules for each domain. At the highest level, the fill rules specify the reporting conditions for a given domain; that is, what information in the document is to be extracted and coded as part of the template. These reporting conditions correspond to the general goals of the extraction task. For example, the fill rules document for the JV domain defines what information in a text constitutes sufficient evidence to report a joint venture. Of note is that the conditions enumerated in the fill rules were determined from the document corpus and refined through the actual application of (earlier versions of) the fill rules to the corpus. At a more specific level, the fill rules delineate the conditions for instantiating an object, object by object, and for filling a slot, slot by slot. At the object and dot levels, the rules specify (1) what kind of evidence in the text is required for instantiation or fills and what, if anything, can be inferred, (2) the formatting conditions for data representation, and (3) the semantics of the data elements. Examples are often provided to highlight any one of these aspects.</Paragraph>
    <Paragraph position="1"> The fill rules served as guidelines for two very different sets of users--the analysts and the system developers.</Paragraph>
    <Paragraph position="2"> Since the evolution of a Fill Rule document was driven to a large extent by its application to a text corpus, the analysts were key contributors to the fill rules in that they applied the rules and in so doing identified discrepancies, omissions, and exceptions to the rules. System developers, on the other hand, were mainly &amp;quot;consumers&amp;quot; of the rules, even though the TIPSTER participants did provide substantial input to the fill rules through questions and comments. Although reporting conditions as well as object and slot specifications need to be implemented in the extraction systems, the developers of those systems also relied on the text corpus itself and analyst-filled templates to direct development.</Paragraph>
    <Paragraph position="3"> In support of the fill rules document, other specialized documents were also provided, for example, expanding on the definition of a joint venture, or on the semantics of representing time expressions. programs, and the test sets were used to measure system wrformance at six-month intervals (see above under TEM-</Paragraph>
  </Section>
  <Section position="6" start_page="135" end_page="135" type="metho">
    <SectionTitle>
5. OTHER SUPPORTING MATERIAL
</SectionTitle>
    <Paragraph position="0"> The Government also supplied on-line supporting materials to the analysts and the TIPSTERIMUC-5 participants. In many cases, this material was used to regularize or normalize the template fills. For example, it was necessary to use the English language Gazetteer to regularize geographic locations. Compiled from a variety of sources, this resource provides place names for more than 240,000 locations around the world. For example, Baltimore is identified as a CITY, located in the PROVINCE (state) of Maryland, which is in the COUNTRY USA. The entire gazetteer entry for a location is used as the normalized fill for locational information in the template. Due to the small number of on-line geographic resources available for Japanese, a much more limited version of a Japanese gazetteer was manually produced by one of the Japanese analysts, with entries for all of the countries in the world, detailed listings for Japanese provinces, U. S. states, major cities for both countries, and other major cities worldwide that appeared in the JJV corpus. The Japanese language Gazetteer contains 1882 different locations.</Paragraph>
    <Paragraph position="1"> In the Joint Venture domain, the reporting of the products or business of the joint venture included classifying the product or service using the Standard Industrial Classification Manual compiled by the U. S. Office of Management and Budget. This resource contains a hierarchical classification of all the industry or business types in the U. S., for example, avocado farms, electric popcorn popper sales, management consulting. The template-filling task required that products or services be coded as a two-digit classification representing the second level in the hierarchy.</Paragraph>
    <Paragraph position="2"> Other supporting resources for fill regularization include lists of currency names and abbreviations (e.g., the Dutch guilder is abbreviated NLG), lists of corporate abbreviations (e.g., Inc, GMBH, and Ltd.) along with lists of countries where those abbreviations are typically used, and nationality adjectives (e.g., Iraqi, Irish). An additional set of resources was provided to system developers to assist in the extraction task, for example, lists of people's first names.</Paragraph>
    <Paragraph position="3"> All of these resources have been made available to the research community through the Consortium for Lexical Research at New Mexico State University.</Paragraph>
  </Section>
  <Section position="7" start_page="135" end_page="135" type="metho">
    <SectionTitle>
6. DATA PREPARATION
</SectionTitle>
    <Paragraph position="0"> The goal of data preparation was to have human analysts produce sets of development and test templates for each of the four corpora. The development templates served as models for system developers in the TIPSTER and MUC PLATE CORPORA). For each of the four languageldomain pairs, a group of experienced analysts was hired. These analysts met regularly over the course of 12 - 21 months (depending on the domain) to discuss domain and language-specific issues, iron out differences, and provide input to the fill rules, which evolved over time. The human analysts used a window-based tool for Sun Microsystems worksta-.</Paragraph>
    <Paragraph position="1"> tions, developed for the template-filling task by New Mexico State University's Computing Research Laboratory. One additional sub-task undertaken as part of the data preparation was the establishment of a performance baseline by measuring the performance of human analysts against each other and against the final &amp;quot;correct&amp;quot; version of various templates (see Table 2 below; for more &amp;tail, see also &amp;quot;Comparing Human and Machine Performance for Natural Language Information Extraction: Results from the Tipster Text Evaluation&amp;quot; in this volume).</Paragraph>
    <Paragraph position="2"> Eleven of the nineteen analysts which comprised the four teams were hied by the Institute for Defense Analyses (IDA) in Virginia; additional analysts from various Government facilities joined these teams. The Government technical management team (including the authors) led the effort to specify the domain, the definition of the templates, and the development of fill rulesfill rules and other supporting materials, in addition to directing IDA, which was responsible for tracking template production and delivering prepared materials to the contractor sites, among other tasks.</Paragraph>
    <Paragraph position="3"> In order to ensure maximal consistency and correctness in the analyst-produced keys, a variety of template-filling schemes were tried. Essentially, the schemes used different degrees of redundancy in producing each filled template, then used different methods to compare those template versions and to produce one final &amp;quot;most correct&amp;quot; version. Table 2 summarizes the different strategies that were tried. Most templates were produced using AB+B or AB+C; JME was entirely produced using A+A. For the other three corpora, the A, B, C, and D positions were rotated among the analysts. Even though redundant coding and checking methods were utilized, the templates that were produced were not perfect; anomalies found by system developers were reviewed and changes were incorporated into the templates as appropriate.</Paragraph>
  </Section>
  <Section position="8" start_page="135" end_page="139" type="metho">
    <SectionTitle>
7. TEMPLATE-FILLING STRATEGIES
</SectionTitle>
    <Paragraph position="0"> The methodology used by the human analysts in filling templates was studied during the course of the task, partly to drive redesign of the tools and documentation to support the analysts' efforts. Although available resources did not permit extensive cognitive study of the mechanisms  One analyst codes, then checks it in a separate pass at a later time One analyst codes template, another checks it Two analysts independently produce codings, then one of them reviews both and produces composite version Two analysts independently produce codings, then a third analyst reviews those two and produces composite version.</Paragraph>
    <Paragraph position="1"> Each of four analysts produces coding independently, then final version produced by entire committee Each of four analysts produces coding independently, then final version produced by the fifth person used by analysts, we did make some general observations about the strategies used by analysts.</Paragraph>
    <Paragraph position="2"> A variety of approaches were used by the human analysts in filling out the templates. What follows is a characterization of the different strategies used by the five Japanese joint venture analysts (referred to as Analysts A, B, C, D and E) in analyzing the documents and filling out the corresponding templates.</Paragraph>
    <Paragraph position="3"> The analysts' task can be divided into two parts: a start-up procedure and the actual template filling process, using the on-line tool. The start-up procedure includes both reading the text, and annotating a hard copy of the document. The template-filling process addresses the order in which the analysts actually filled out the objects and slots that represented the various pieces of information to be extracted from the text.</Paragraph>
    <Paragraph position="4"> For the start-up procedure, three distinct approaches were identified. Scheme 1, used by two analysts, is characterized by minimal marking of the hard-copy text before starting to code the template using the on-line tool. Analyst B would read the article twice through, then underline and label just the tie ups and entities before going to the tool. Analyst D would read and simultaneously underline entities and place check marks by other pertinent data; then he would begin coding.</Paragraph>
    <Paragraph position="5"> In Scheme 2, also used by two analysts, a more detailed annotation of the hard-copy text was made. Analyst E would read through the hard-copy text and simultaneously underline and number entities, circle and number tie ups, and make comments, such as &amp;quot;El alias,&amp;quot; &amp;quot;E2 official&amp;quot; (for alias or official~associated with a particular entity). Moreover, this analyst would draw links between related pieces of information in the text, and would outline in the  margins more complex objects, such as ACTIVITY, OWN-ERSHIP, and REVENUE. After this process was complete, the coding would begin. Analyst C's approach was similarly detailed, the only difference being that she would label all pertinent information using color-coded highlighters, e.g., green for ENTITYS, yellow for product/service strings, blue for FACILITY and TIME objects.</Paragraph>
    <Paragraph position="6"> The thkd scheme, used only by Analyst A, involved a mixture of initial marking, skimming, initial coding, annotating in detail, and then final coding. This analyst would read the beginning of the article, marking potential entities until a &amp;quot;tie-up verb&amp;quot; was found. Now certain that the article had a valid tie up, she would proceed to skim the remainder of the text, underlining or circling additional pertinent information. At this point, she would use the tool to code the initial portion of the template, i.e., the TIE-UPs, ENTITYs, and ENTITY-RELATIONSHIPS. After this key structure was in place, she would read through the remainder of the text, annotating in detail all potential product/service strings and information about FACILITYs, REVENUE, OWNERSHIP, etc. Finally, the remainder of the template was coded using the tool.</Paragraph>
    <Paragraph position="7"> In the templateTfilling process, a variety of breadth vs. depth-first strategies were used by the analysts. Four of the analysts would completely fill in all information about the first tie up before coding any additional tie ups. Analysts A, B, and E would fill in the TEMPLATE, TIE-UP, ENTITY, and ENTITY-RELATIONSHIP objects first. Then TIME, REVENUE, OWNERSHIP, PERSON and FACILITY objects were instantiated in no particular order. The ACTIVITY and INDUSTRY objects were filled in concurrently, usually last. This procedure was then repeated for additional tie ups. Analyst D followed a complete depth-first strategy for coding each tie up, filling in each slot in turn, so that if a slot pointed to another object, that object would then be filled in completely before proceeding to the next slot in the top level object. A breadth-first strategy for coding was used by Analyst C, who would fill in all tie-up objects and their respective entities first, and then code the remaining information for each tie up.</Paragraph>
    <Paragraph position="8"> These varying strategies for annotating texts and coding templates did not seem to have a significant effect on the quality of the templates produced, and seemed to be a matter of personal preference. However, they give insight into the different ways in which humans approach a particular analytic task, and suggest that on-line analytic tools need to be sufficiently flexible to accommodate the styles of different human users.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML