File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/82/c82-1013_metho.xml

Size: 11,124 bytes

Last Modified: 2025-10-06 14:11:25

<?xml version="1.0" standalone="yes"?>
<Paper uid="C82-1013">
  <Title>FORWARD AND BACKWARD REASONING IN AUTOMATIC ABSTRACTING</Title>
  <Section position="3" start_page="83" end_page="83" type="metho">
    <SectionTitle>
SPECIFICATIONSAND BASIC METHODOLOGIES
</SectionTitle>
    <Paragraph position="0"> In defining SUSY's specifications we have tried to implement at a certain level of abstraction an important human feature: the capability to generate summaries of different content and extent depending on the user's goals. The system is therefore able to process a text following the two principles of variable-length processing and of user-taylored abstracting. With variable-length processing we mean the Capability to generate, starting from the same text. summaries of different length, complexity, and level of abstraction depending on the user's requirements. With user-taylored abstractingwe mean the capability to generate s~ries of different content depending on the user's goals and needs.</Paragraph>
    <Paragraph position="1"> Together with the input text, SUSY can therefore receive in input the user's requirements describing with more or less details the organization, content, and extent of the output summary. This is done through a summary schema which can be interactively supplied at the beginning of the session. The user can also provide the system with a text schema which is constituted by a set of suggestions on how the input text can be interpreted. The text schema has a twofold motivation: to help the system in capturing from the input text only the most relevant parts, and to increase s~rizing effectiveness.</Paragraph>
    <Paragraph position="2"> Turning now our attention to the methodological aspects of SUSY. we notice that.</Paragraph>
    <Paragraph position="3"> in general, the surmnsrizing activity can be performed in two distinct and complementary ways. The first one. or meaning-based, is grounded on the comprehension of the text to be summarized: in this case the summarizer has to capture the most important information contained in the text. The second possible way is structure-based and it does not rely on the meaning of the text but rather on its structure: the summary is obtained by eliminating, without understanding, parts of the text (for example adjectives, relative sentences.</Paragraph>
    <Paragraph position="4"> etc.) which a priori are considered less relevant. Both these ways can be combined with the two basic methodologies we have conceived for the system, i.e.</Paragraph>
    <Paragraph position="5"> forward and backward processing.</Paragraph>
    <Paragraph position="6"> With the term forward processing we mean the capability to understand the whole natural language text and to produce in output, possibly through the iterative application of summarizing rules, the desired summary. This is clearly a bottom-up approach which constantly focuses on the input text. In backward processing, on the other hand. the focus is on the s~ry schema. The system works now top-down, searching for those parts of the text that can be utilized to build up the summary according to the specifications contained in the summary schema. In the SUSY sistem we have chosen to implement both forward and backward processing within a meaning-based approach.</Paragraph>
    <Paragraph position="7"> SYSTEM ARCHITECTURE AND BASIC ALGORITHMS The architecture of the system is organized in two main parts: the first one is devoted to collect the user's requirements and suggestions and to perform a preprocessing activity on them, the second one implements the actual parsing.</Paragraph>
    <Paragraph position="8"> summarizing, and generation activities.</Paragraph>
    <Paragraph position="9"> The first par~ of the system constitutes an interactive interface centered around a main module called schema builder. This module is devoted to engage a</Paragraph>
    <Paragraph position="11"/>
  </Section>
  <Section position="4" start_page="83" end_page="83" type="metho">
    <SectionTitle>
FORWARD AND BACKWARD REASONING IN AUTOMATIC ABSTRACTING &amp;quot; . 85
</SectionTitle>
    <Paragraph position="0"> bounded scope dialogue with the user in order to collect his suggestions about the structure and content of the texts to be su~mmrized, and his requirements on the summary to be generated. This information is embedded in two different frameworks called working_ text schema and working summmar~, schema which contain the user's suggestions and requirements, respectively. The schemas will constitute a fundamental input for the following phases of the system operation.</Paragraph>
    <Paragraph position="1"> The working schemas are defined by the user. under the continuous guidance of the schema builder, through three different activities: - choosin~ the most appropriate schema from a library of basic text and summary schemas or from a library of working text and summary schemas which contain the schemas utilised in previous surm~arizing sessions; - tuning a selected schema by assigning (or reassigning) same parameters contained in it; - defining a fully new (basic) schema.</Paragraph>
    <Paragraph position="2"> It is understood that working schemas are not requested to be always defined at the same level of detail and completeness; they are allowed to embed more or less information according to the adequacy and richness of the specifications supplied by the user. For both text and summary schemas there exist default values to be utilized when the user is unable or unwilling to supply its own specifications.</Paragraph>
    <Paragraph position="3"> The second part of the system is devoted to the parsing, surmnarizing, and generation activities. These are conceivedin SUSY as three sequential steps which conlnunicate through precisely defined data interfaces representing intermediate results of the processing.</Paragraph>
    <Paragraph position="4"> The parser constructs the internal representation of the input text on which the summarizer will afterwards perform its activity. The operation of this module is based on a semantics-directed parsing algorithm which aims to supply a full understanding of the input text along the following two main lines: - the text is parsed in a uniform way. independently of any expectation that could be possibly made (by considering the current working schemas) about the relevance of the different parts of the text in relation with the summary to be produced; - the parsing is performed at a generally high level of abstraction, without decomposing objects into very elementary semantic primitives (Schank. 1975) but only considering the basic attributes and relations which are necessary for the summarizing task.</Paragraph>
    <Paragraph position="5"> The semantics directed parsing algorithm utilises two kinds of information: the elementary knowledge about words and simple constructs contained in the vocabulary, and a set of semantic rules that specify the basic properties and relations of the elementary semantic entities which are supposed to play a role in the application domain in which the system operates (Guida and Tasso. 1982). The internal representation constructed by the parser shares many features with that proposed by Kintsch (Kintsch, 1974; Kintsch and van Dijk. 1978) and is constituted by a sequence of labelled linear propositions each one conveying a unit of information. Every proposition is composed by a predicate with one or more arguments. Predicates and arguments can be considered as concepts or types to which the words in the input text (tokens) refer. The same type may be 86' D. FUM, G. GU1DA and C. TASSO instantiated by different tokens which are therefore considered as synonlms. Arguments can be types or labels of propositions and~ in any case, they play .precise semantic roles (agent. object, patient etc.). Every predicate imposes some constraints (linguistic or derived from the world knowledge possessed by the system) on the number and nature of its arguments. The proposions are connected to each other through shared terms in such a way to represent an actual network structure.</Paragraph>
    <Paragraph position="6"> The activity of the summarizer has been split, according to the basic methodology illustrated in the previous section, in two sequential steps: a forward one performed by the weighter and a backward one implemented by the selector. The weighter is devoted to organize the internal representation, which is originally a flat and homogeneous network, into a structured framework in which the different levels of relevance and detail of the single propositions are clearly defined. This is obtained by assigning an integer weight to each proposition in such a way to generate a weighted network called weighted representation. The weighter utilizes for its operation the working text schema and a set of general purpose weighting rules. The selector is devoted to prune the ~ighted internal representation in such a way to obtain the selected representation i.e. the internal representation of the desired sunmmry. It takes into account the working summary schema and operates through a set of general purpose selecting rules. The pruning it performs is generally not uniform with respect to the weights attached to the weighted representation, but it is biased and tuned by the requirements contained in the sun~nary schema.</Paragraph>
    <Paragraph position="7"> It is easy to recognize that weighting is indeed a forward activity which mainly focuses on the input text. while selecting represents a backward process which is generally directed by the consideration of the summary to be generated. Let us outline that the completeness and depth of the weighting and selecting activities strongly depend on the quality and richness of the text and summary schemas, respectively. Generally. these steps are not equally balanced and. in some cases, one of themmay even be nearly void. as text schema or summary schema may be almost empty or even missed. In such cases we obtain a pure forward or backward strategy.</Paragraph>
    <Paragraph position="8"> The last step of the system operation is the actual generation of natural language summary that is performed by the generator. Its activity is organised in two phases: - retrieval from the input text of the basic linguistic elements (words.</Paragraph>
    <Paragraph position="9"> phrases, whole sentences etc.) necessary to compose the summary; - appropriate assembly of these elements into a correct and acceptable text.</Paragraph>
    <Paragraph position="10"> In the second phase it utilizes a set of sentence models which supply the most basic and usual rules for constructing correct sentences in a simple and plain style.</Paragraph>
  </Section>
  <Section position="5" start_page="83" end_page="83" type="metho">
    <SectionTitle>
AN EXAMPLE
</SectionTitle>
    <Paragraph position="0"> Owing to space restrictions we present in this section only a short working example of SUSY's performance, focusing on the most relevant features of the internal representation and of the weighting and selecting activities.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML