File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/99/w99-0204_metho.xml

Size: 21,738 bytes

Last Modified: 2025-10-06 14:15:25

<?xml version="1.0" standalone="yes"?>
<Paper uid="W99-0204">
  <Title>Automatic Slide Presentation from Semantically Annotated Documents</Title>
  <Section position="3" start_page="0" end_page="25" type="metho">
    <SectionTitle>
2 The GDA Tagset
</SectionTitle>
    <Paragraph position="0"> GDA is a project to make WWW texts machineunderstandable on the basis of a. linguistic tag set, and to develop applications such as content-based presentation, retrieval, question-answering, summarization, and translation with much higher quality than before. GDA thus proposes an integrated global platform for electronic content authoring, presentation, and reuse. The GDA tagset 1 is based on XML, and designed as compatible as possible with HTML, and TEI 2, etc., incorporatlhttp ://~w. etl. go. jp/et I/nl/GDA/t agset, html  ing insights from EAGLES s, Penn TreeBank \[Marcus et al., 1993\], and so forth.</Paragraph>
    <Paragraph position="1"> Described below is a minimal outline of the GDA tagset necessary for the rest of the discussion. Parse-tree bracketing, semantic relation, and coreference are essential for slide presentation, as with many other applications such as translation. Further details, concerning coordination, scoping, illocutionary act, and so on, are omitted.</Paragraph>
    <Section position="1" start_page="25" end_page="25" type="sub_section">
      <SectionTitle>
2.1 Parse-Tree Bracketing
</SectionTitle>
      <Paragraph position="0"> As the primary purpose of GDA tagging is to encode semantic structure, syntactic annotation is exploited only as far as it contributes to semantic encoding. Also, syntactic tags are designed to simplify syntactic annotation by minimizing the number of tags and accordingly the depth of embedding among them.</Paragraph>
      <Paragraph position="1"> An example of a GDA-tagged sentence is shown in  &lt;adp&gt; stand for noun phrase, verb, and adnominal or adverbial phrase.</Paragraph>
      <Paragraph position="2"> &lt;su&gt; and the tags whose name end with 'p' (such as &lt;adp&gt; and &lt;vp&gt;) are called phrasal tags. In a sentence, an element (a text span enclosed in a begin tag and the corresponding end tag) is usually a syntactic constituent. The elements enclosed in phrasal tags are phrasal elements, which cannot be the head of larger elements. So in Figure 1 'flies' is specified to be the head of the &lt;su&gt; element and 'like' the head of the &lt;adp&gt; element.</Paragraph>
    </Section>
    <Section position="2" start_page="25" end_page="25" type="sub_section">
      <SectionTitle>
2.2 Semantic Relation
</SectionTitle>
      <Paragraph position="0"> The tel attribute encodes a relationship in which the current element stands with respect to the element that it syntactically depends on. Its value represents a binary relation, which may be a grammatical function such as SUBJECT, a thematic role such as AGENT, PATIENT, RE-CIPIENT, or a rhetorical relation such as CAUSE, CONCES-SION, and ELABORATION. Grammatical functions are used to encode semantic relation assuming that a dictionary is availableby which to associate grammatical functions with thematic roles for lexical items such as verbs. Thematic roles and rhetorical relations are also conflated, because the distinction between them is often vague. For instance, CONCESSION may be both intrasentential and intersentential relation.</Paragraph>
    </Section>
    <Section position="3" start_page="25" end_page="25" type="sub_section">
      <SectionTitle>
2.3 Coreference
</SectionTitle>
      <Paragraph position="0"> As discussed later, coreferences play a major role in slide presentation, id, eq, ctp, sub and sup attributes are mainly used to encode coreferences. Each element may have an identifier as the value for the id attribute. Coreferent expression should have the eq attribute with its antecedent's id value. An example follows: &lt;np id=&amp;quot;jO&amp;quot;&gt;John&lt;/np&gt; beats &lt;adp eq=&amp;quot;jO&amp;quot;&gt;his&lt;/adp&gt; dog.</Paragraph>
      <Paragraph position="1"> When the shared semantic content is not the referent but the type (kind,set,etc) of referents, the ctp attribute is used.</Paragraph>
      <Paragraph position="2"> You bought &lt;np id=&amp;quot;cl&amp;quot;&gt;a car&lt;/np&gt;.</Paragraph>
      <Paragraph position="3"> I bought &lt;np ctp=&amp;quot;cl&amp;quot;&gt;one&lt;/np&gt;, too.</Paragraph>
      <Paragraph position="4"> The values for the rel attribute also function as attributes, called relational attributes. A zero anaphora is encoded by a relational attribute.</Paragraph>
      <Paragraph position="5"> Tom visited &lt;np id=&amp;quot;ml&amp;quot;&gt;Mary&lt;/np&gt;.</Paragraph>
      <Paragraph position="6"> He had &lt;v iob=&amp;quot;ml&amp;quot;&gt;brought&lt;/v&gt; a present.</Paragraph>
      <Paragraph position="7"> iob=&amp;quot;ml&amp;quot; means that the indirect object of brought is element ml, that is, Mary.</Paragraph>
      <Paragraph position="8"> Other relational attributes in this connection include sub and sup. sub represents subset, part, or element. An example is: She has &lt;np id=&amp;quot;bl&amp;quot;&gt;many books&lt;/np&gt;.</Paragraph>
      <Paragraph position="9"> &lt;namep sub=&amp;quot;bl&amp;quot;&gt; ' ' Alice ~ s Adventures in Wonderland ~ ~&lt;/namep&gt; is her favorite.</Paragraph>
      <Paragraph position="10"> sup is the inverse of sub, i.e., includer of any sort, which is superset as to subset, whole as to part, or set as to element.</Paragraph>
    </Section>
  </Section>
  <Section position="4" start_page="25" end_page="28" type="metho">
    <SectionTitle>
3 Making Slide Show
</SectionTitle>
    <Paragraph position="0"> We have developed a system which generates slide shows from GDA-tagged documents. Our method for slide presentation consists of two aspects. The first is to detect topics in the given document. The second aspect is to generate slides for the topics and organize them to a slide show. The latter employs some language-dependent heuristics. But neither aspect uses any heuristics dependent on the domain and/or style of documents. So our method is potentially applicable to any GDA-tagged documents.</Paragraph>
    <Section position="1" start_page="25" end_page="26" type="sub_section">
      <SectionTitle>
3.1 Topic Detection
</SectionTitle>
      <Paragraph position="0"> Topics are often represented by important words and/or phrases in the documents. A traditional method for topic identification is to use word/phrase-occurrence frequencies to extract such expressions. Such a method is not adequate for extracting topics, however, due to the following reasons:  1. A word is often too short to fully reresent a topic. 2. A topic is often represented by a variety of expressions. null  For example, if we count the frequencies of the words in an article of the Wall Street Journal, which is in Figure 2, discard the words whose frequencies are less than two, and drop stop words, then we get</Paragraph>
      <Paragraph position="2"> where the numbers are the frequencies. From this list, we know the article is about PCs. But it is doubtful that the list distinguishes the article from other articles which also describe PCs.</Paragraph>
      <Paragraph position="3"> To remedy these problems, we may extract word bi-grams in addition to word unigrams or use a stemmer to normalize expressions. But these are not fundamental solutions.</Paragraph>
      <Paragraph position="4"> Instead we use semantic dependencies and coreferences for identifying topics. First we collect syntactic subjects and classify them according to their referents, and then discard the classes consisting of less than two elements. Next, we choose representative expressions from these classes and regard them as topics. A representative expression of a class is the element which is assigned the id attribute related with the class unless the element is elaborated by another element. If it is elaborated, then the elaborating expression is selected as representative. For example, we can extract the following four topics from the WSJ article.</Paragraph>
      <Paragraph position="5">  where the numbers are the sizes of the classes. Note that &amp;quot;the Apple II, Commodore Pet and Tandy TRS&amp;quot; does not have an id attribute because it is a coreference expression whose antecedent is &amp;quot;THREE COMPUTERS THAT CHANGED the face of personal computing.&amp;quot; Nevertheless it is selected as a topic because it elaborates its antecedent. Note also that &amp;quot;many pioneer PC contributors&amp;quot; is not a subject but it is selected as the representative expression of &amp;quot;William Gates and Paul Allen,&amp;quot; &amp;quot;Gates,&amp;quot; &amp;quot;Alan F. Shugart,&amp;quot;and &amp;quot;Dennis Hayes and Dale Heatherington&amp;quot; because it has an id attribute and is pointed by the other expressions with sub relation.</Paragraph>
      <Paragraph position="6"> We believe that the expressions extracted by using syntactic and coreference information is much more appropriate for topics than the ones based on word frequencies. It is, however, a future work to confirm it experimentally.</Paragraph>
    </Section>
    <Section position="2" start_page="26" end_page="27" type="sub_section">
      <SectionTitle>
Topic Selection
</SectionTitle>
      <Paragraph position="0"> Frequency is not enough to distinguish the importances of topics (words and/or phrases) because different topics often have the same frequency. So we use a sort of spreading activation \[Nagao and Hasida, 1998\] to calculate the importance of elements. A GDA-tagged document is regarded as a network in which nodes correspond to GDA elements and links represent the syntactic During its centennial year, The Wall Street Journal will report events of the past century that stand as milestones of American business history. THREE COMPUTERS THAT CHANGED the face of personal computing were launched in 1977. That year the Apple II, Commodore Pet and Tandy TRS came to market. The computers were crude by today's standards. Apple II owners, for example, had to use their television sets as screens and stored data on audiocassettes.</Paragraph>
      <Paragraph position="1"> But Apple II was a major advance from Apple I, which was built in a garage by Stephen Wozniak and Steven Jobs for hobbyists such as the Homebrew Computer Club. In addition, the Apple II was an affordable $1,298. Crude as they were, these early PCs triggered explosive product development in desktop models for the home and office. Big mainframe computers for business had been around for years.</Paragraph>
      <Paragraph position="2"> But the new 1977 PCs - unlike earlier built-from-kit types such as the Altair, Sol and IMSAI - had keyboards and could store about two pages of data in their memories. Current PCs are more than 50 times faster and have memory capacity 500 times greater than their 1977 counterparts. There were many pioneer PC contributors. William Gates and Paul Allen in 1975 developed an early language-housekeeper system for PCs, and Gates became an industry billionaire six years after IBM adapted one of these versions in 1981.</Paragraph>
      <Paragraph position="3"> Alan F. Shugart, currently chairman of Seagate Technology, led the team that developed the disk drives for PCs. Dennis Hayes and Dale Heatherington, two Atlanta engineers, were co-developers of the internal modems that allow PCs to share data via the telephone. IBM, the world leader in computers, didn't offer its first PC until August 1981 as many other companies entered the market. Today, PC shipments annually total some $38.3 billion world-wide.</Paragraph>
      <Paragraph position="4">  dominance and semantic relationships described before.</Paragraph>
      <Paragraph position="5"> That is, this network is the tree of GDA elements plus cross-reference links among the nodes therein. Spreading activation applies to this network. It is performed respecting the condition that two elements should have the same activation value if either they are coreferent or one of them is a syntactic head of the other.</Paragraph>
      <Paragraph position="6"> When we apply spreading activation to the WSJ article, we get the following activation values for the topics: the Apple II , Commodore Pet and Tandy * (9.61)</Paragraph>
      <Paragraph position="8"/>
    </Section>
    <Section position="3" start_page="27" end_page="27" type="sub_section">
      <SectionTitle>
Apple II
</SectionTitle>
      <Paragraph position="0"> We can pick up the top two as the most important topics which will be presented in the slide show if we discard the topics whose activation values are smaller than a half of that of the top topic. We can also display this whole list to the audience so that he/she/they can choose topics to be presented in the rest of the slide show.</Paragraph>
    </Section>
    <Section position="4" start_page="27" end_page="28" type="sub_section">
      <SectionTitle>
3.2 Slide Generation
</SectionTitle>
      <Paragraph position="0"> A slide show is created by composing a slide for each topic selected as discussed above. In the current implementation of the slide presentation system, each slide is basically an itemized summary of the segment concerning the topic.</Paragraph>
      <Paragraph position="1"> The initial slide may be a table of contents of the whole slide show, which is compiled by listing the topics. Each slide in the main body of the presentation is composed by following the steps below. Here a topical element is an  GDA element linked with the topic via the eq, ctp, sub, or sup relation. A topical element which is the subject of a whole sentence is called a topical subject.</Paragraph>
      <Paragraph position="2"> 1. Let the topic be the heading of the slide.</Paragraph>
      <Paragraph position="3"> 2. Extract important sentences which contain topical subjects.</Paragraph>
      <Paragraph position="4"> 3. Remove redundant sentences, such as one elaborated by another extracted sentence, where elaboration is encoded by the ela relation.</Paragraph>
      <Paragraph position="5"> 4. Itemize the remaining sentences by the following heuristics, among many others.</Paragraph>
      <Paragraph position="6"> (a) Prune unimportant expressions such as some (typically unrestrictive) relative clauses and appositive phrases.</Paragraph>
      <Paragraph position="7"> (b) Remove the topical subjects linked with the topic through the eq or ctp relation.</Paragraph>
      <Paragraph position="8"> (c) Pronominalize non-subject topical elements linked with the topic through the eq or ctp relation.</Paragraph>
      <Paragraph position="9"> (d) Emphasize the topical elements linked with the topic through the sub or sup relation.</Paragraph>
      <Paragraph position="10"> (e) Replace non-topical anaphoric elements with their antecedents.</Paragraph>
      <Paragraph position="11"> if) Move the elements preceding the removed topical subjects to the end of the sentences.</Paragraph>
      <Paragraph position="12"> (g) Decompose coordinate structures whose con null junctions are and, as well as, not only ~ but also, etc. into separate items.</Paragraph>
      <Paragraph position="13"> Heuristics (a) through (g) are specific to English, but it is straightforward to adapt them to other languages. The above WSJ article eventually gives rise to the three slides in Figure 3, Figure 4, and Figure 5.</Paragraph>
      <Paragraph position="14">  The first slide in Figure 3 is the table of contents. The second slide is titled by the first topic in the article, followed by a list of items. To compose this list, initially the following sentences are picked up which talk about the topic.</Paragraph>
      <Paragraph position="15">  1. THREE COMPUTERS THAT CHANGED the face of personal computing were launched in 1977.</Paragraph>
      <Paragraph position="16"> 2. That year the Apple II, Commodore Pet and Tandy TRS came to market.</Paragraph>
      <Paragraph position="17"> 3. The computers were crude by today's standards. 4. Crude as they were, these early PCs triggered ex null plosive product development in desktop models for the home and office.</Paragraph>
      <Paragraph position="18">  5. But the new 1977 PCs - unlike earlier built-from-kit types such as the Altair, Sol and IMSAI - had keyboards and could store about two pages of data in their memories.</Paragraph>
      <Paragraph position="19"> The first sentence is abandoned because it is elaborated by the second. In the other sentences, unnecessary subexpressions are pruned off due to (a) and the references to the topic are replaced by C/ due to (b), as follows:  1. That year C/ came to market.</Paragraph>
      <Paragraph position="20"> 2. C/ were crude.</Paragraph>
      <Paragraph position="21"> 3. C/ triggered explosive product development. 4. C/ had keyboards and could store about two pages  of data.</Paragraph>
      <Paragraph position="22"> The first sentence above is then paraphrased by replacing &amp;quot;that year&amp;quot; with &amp;quot;in 1977&amp;quot; due to (e) and moving it at the end due to (f). The coordinate structure in the last sentence is decomposed into two list items due to (g). The final result is the slide shown in Figure 4. The third slide is composed in essentially the same way as the second, except that the topical subjects are emphasized due to (d) as shown in Figure 5. Further details are omitted.</Paragraph>
      <Paragraph position="23"> From preliminary experiments, we found that the above heuristics work fine for many cases. But in some cases they break down. For example, applying heuristic (a) to &amp;quot;The Wall Street Journal will report events of the past century that stand as milestones of American business history.&amp;quot; produces &amp;quot;The Wall Street Journal will report events,&amp;quot; which is not appropriate because the resulting sentence lacks the information necessary to describe what event the WSJ is going to report. Such a problem may be avoided if there are pragmatic tags to encode which parts of the document somehow convey new information.</Paragraph>
    </Section>
    <Section position="5" start_page="28" end_page="28" type="sub_section">
      <SectionTitle>
3.3 Dynamic Adaptation
</SectionTitle>
      <Paragraph position="0"> Under the framework described so far, it is straightforward to dynamically adapt a presentation to the audience's requests. This is done by reflecting interactions with the audience in the evaluation of importance and topic selection. This adaptation of importance evaluation and topic selection leads to reorganization of the presentation.</Paragraph>
      <Paragraph position="1"> The current presentation system deals with a simple type of interaction which allows the audience to issue questions about parts of the document. This is done in two ways, one by clicking on the screen and the other by typing on the keyboard. A click on a point in a slide is to select the smallest element containing that point. A further click on .the selected element is to select its parent element, and so forth. Having specified a part of the document, whether by clicking or typing, the audience can then request an explanation about it. A new slide is made and shown on the fly if the original document contains more information (absent in the present slide) about that phrase. The remaining part of the presentation, if any, incorporates such interaction by evaluating the specified phrase more importance than otherwise.</Paragraph>
      <Paragraph position="2"> For instance, suppose the audience asks about 'IBM' at some point in the slide show from Figure 3 to Figure 5. Then a slide shown in Figure 6 will be composed IBM * adapted an early language-housekeeper system in 1981.</Paragraph>
      <Paragraph position="3"> * did n't offer its first PC until August 1981.</Paragraph>
    </Section>
  </Section>
  <Section position="5" start_page="28" end_page="29" type="metho">
    <SectionTitle>
4 Concluding Remarks
</SectionTitle>
    <Paragraph position="0"> We have discussed automatic generation of slide presentations from semantically annotated documents. The reported presentation system first detects important topics in the given document and then creates a slide for each topic. Coreferences play a central role in both topic identification and paraphrasing summarization.</Paragraph>
    <Paragraph position="1"> The presentation can be dynamically customized by reflecting the interaction with the audience in topic selection and importance evaluation. Since the GDA tagset is independent of the domain andstyle of documents and also applicable to diverse natural languages, the reported system is domain/style-free and easy to adapt to different languages as well.</Paragraph>
    <Paragraph position="2"> There is no established formal method for evaluating a technology such as slide presentation. We are hence attempting to evaluate partial aspects of the reported  method, such as topic selection and paraphrasing. A more synthetic evaluation is a future work.</Paragraph>
    <Paragraph position="3"> There are several avenues along which to improve or extend the reported system. First, it should be easy to incorporate figures and tables into the slides from the original document. These non-textual materials can also be treated as GDA elements and processed in the same way as text elements with respect to importance evaluation. Second, textual materials could often be rendered visually more perspicuous than a mere list of items. For instance, some sorts of textual content could be naturally depicted by a graph with labeled nodes and arrows, on the basis of spatial metaphors. Third, not just subjecthood but also other grammatical functions and anaphoricity of the relevant expressions could be used to identify topics. The intuitions behind centering theory \[Grosz et al., 1995\] may be useful here. Finally, more sophisticated types of interaction than described above are desirable and feasible, including question answering.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML