File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/82/c82-1049_metho.xml

Size: 10,771 bytes

Last Modified: 2025-10-06 14:11:31

<?xml version="1.0" standalone="yes"?>
<Paper uid="C82-1049">
  <Title>A MESSAGE-PASSING CONTROL STRUCTURE FOR TEXT UNDERSTANDING</Title>
  <Section position="4" start_page="0" end_page="0" type="metho">
    <SectionTitle>
3.0 THE APPLICATION
</SectionTitle>
    <Paragraph position="0"> We are using working abstracts of descriptions of Texas Instruments' issued patents. These are written in a restricted style by by the attorneys and the topics are limited to solid-state micro-electronic devices. Thus we are able to use naturally occurring data without immediately confronting the problems of incomplete syntactic and conceptual coverage that would be encountered in many other domains. An example is: A modulator comprising two transistors each having collector, emitter and base electrodes, means for applying a direct voltage across said emitter electrodes, a center-tapped source of alternating signals connected between said base electrodes, said collector electrodes being connected together and to the center tap of said source. A load impedance connected between said collector electrodes and said emitter electrode of one of said transistors, and a variable resistor connected between the base electrode and the emitter electrode of said one transistor.</Paragraph>
    <Paragraph position="1"> The interaction of embedding and conjunction gives a high degree of syntactic ambiguity to the texts. The texts can also be ungrammatical, whence the desire to be building the conceptual representation in parallel with the syntactic analysis in order that some meaning will be extracted from the text even when the syntactic analysis is not completed.</Paragraph>
    <Paragraph position="2"> The goal of the project is to build a conceptual representation for the text, then add retrieval capabilities that will be more flexible than a word-matching scheme.</Paragraph>
  </Section>
  <Section position="5" start_page="0" end_page="0" type="metho">
    <SectionTitle>
4.0 THE SYSTEM
</SectionTitle>
    <Paragraph position="0"> The objects of our system correspond to the organizing principles of the components. In syntax we have constituency objects that can take grammar rules and try to match them against input. In semantics we have taxonomy objects, case-frame objects, causal-chain objects, meta objects (that handle a form of lambda-abstraction), etc.</Paragraph>
    <Paragraph position="1"> Small &amp; Rieger (1981) also have an object viewpoint of language analysis.</Paragraph>
    <Paragraph position="2"> However, in their scheme each word is an object, motivated by their view that &amp;quot;human knowledge about language is organized primarily as knowledge about words rather than as knowledge about rules&amp;quot; (p.l). Our system is organized around rules.</Paragraph>
    <Paragraph position="3"> Objects in different components do not necessarily have the same vocabulary: in syntax there are words and phrase structure and semantics has concepts and relations; accordingly there is a translation object through which messages pass.</Paragraph>
    <Paragraph position="4"> A CONTROL STRUCTURE FOR TEXT UNDERSTANDING 309 Kornfeld (1979) has given an example of a (pseudo-)parallel communication system that passes information between objects to reduce the respective search spaces; he terms the phenomenon .&amp;quot;combinatorial I__MMplosion&amp;quot;. The interaction between syntax and semantics allows the the conceptual representation of sentence fragments to be built in parallel with the syntactic analysis. Semantic predictions arm fed back to syntax to try to achieve the combinatorial implosion.</Paragraph>
    <Section position="1" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
4.1 The Syntax
</SectionTitle>
      <Paragraph position="0"> The formalism we are using is &amp;quot;local grammar&amp;quot; (Saenz (1980), Ross (1981)) which consists of a context-free phrase structure grammar with augmentations. The augmentations are blocking and percolation rules. For example, the auxiliary rule in our system is</Paragraph>
      <Paragraph position="2"> The structure segments are numbered left to right, starting at 0 for the left-hand-side of the rule. The values of the features &amp;quot;aspect&amp;quot; and &amp;quot;affix&amp;quot; are established in the dictionary for terminal items and percolated up the analysis tree for higher level phrases.</Paragraph>
      <Paragraph position="3"> The parsing algorithm is a modified left-corner routine (Griffiths &amp; Petrick (1965)). The modifications are to use the object environment to produce all parses in parallel and to merge common subparses.</Paragraph>
      <Paragraph position="4">  using the auxiliary rule.of Figure 1. &amp;quot;Category&amp;quot; is the left-hand-side of the structure rule, &amp;quot;part-parse&amp;quot; is the fragment of the right-hand-side so far. matched, and &amp;quot;rule-tail&amp;quot; is the remaining part of the structure rule. &amp;quot;Augmentations&amp;quot; are the percolation and blocking rules. The &amp;quot;goals-list&amp;quot; gives other constituent objects that are awaiting completion of this constituent (there may be several because of merged paths). The word &amp;quot;being&amp;quot; has just been processed as the first segment of the part-parse and the segment counter now</Paragraph>
      <Paragraph position="6"> points to the second position. The object has processes for (a) advancing the analysis with a new input word, (b) for advaacing the analysis when a subparse has been completed (the parse can only be immediately advanced when the next element of the rule-tail is a terminal symbol; otherwise it has to create another object to process a rule expanding the non-terminal category), and (c) for merging with another path.</Paragraph>
    </Section>
    <Section position="2" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
4.2 The Knowledge Base
</SectionTitle>
      <Paragraph position="0"> General knowledge of the domain is represented in a semantic network (Phillips (1978)). The conceptual analyses will be instantiations of general knowledge with novelty introduced from the texts.</Paragraph>
      <Paragraph position="1"> Semantic nets are usually seen as data (e.g., Qillian (1968), Brachman (1978)) with various routines for performing operations such as taxonomy searches, finding paths from node to node, and binding variables in the network. Each node of our network is an object having associated processes dependent on the types of links it has to other nodes.</Paragraph>
      <Paragraph position="2">  Thus in Figure 3, which shows part of our semantic network, links indicate the other nodes to which a node can send a message, as opposed to a physical pointer. A node is actually implemented as a &amp;quot;mix&amp;quot; of objects for the kinds of links it has; thus as new links are added so are new processes. A node need not know anything about the internal format of data in other nodes to get A CONTROL STRUCTURE FOR TEXT UNDERSTANDING 311 information from them. Further, when a message is sent to a node, the sender need not know whether this is a &amp;quot;simple&amp;quot; or &amp;quot;complex&amp;quot; message: a simple message can be answered by the qode itself, a complex message requires the node to send messages to other nodes. Thus a &amp;quot;part-whole&amp;quot; message to establish whether a direct-voltage source can be part of a modulator or part of a transistor, will also use intervening taxonomy, decomposition (meta) links (Figure 3) without this being specified in the original query. A node receiving this message would &amp;quot; pass a similar message to its neighbouring nodes if it cannot itself respond.</Paragraph>
    </Section>
    <Section position="3" start_page="0" end_page="0" type="sub_section">
      <SectionTitle>
4.3 Flow Of Control
</SectionTitle>
      <Paragraph position="0"> Processing of text is initiated from a task specific knowledge object, in this case a &amp;quot;patent knowledge expert&amp;quot; that has an expectation of finding a patent.</Paragraph>
      <Paragraph position="1"> It passes a message to the translator that sees if it has any data that will match this expectation. Since no part of the text has been examined by syntax, nothing can be found. A start message is sent to syntax.</Paragraph>
      <Paragraph position="2"> The translator object should pass predictions to syntax but this does not, in general, seem possible as the realizations of a concept include all possible descriptive references. Thus the translator maintains its list of predictions and, when syntactic constituents are received, matches their translations against the predictions. A match causes a message to be sent to the source of the prediction, which can extend the conceptual representation and produce further predictions.</Paragraph>
      <Paragraph position="3"> When no matches are found, the translator seeks a knowledge structure that corresponds to the syntactic structure. This occurs, for example, when the syntactic objects are confronted with the attachment of the &amp;quot;means for applying ...&amp;quot; phrase in the example given above: is the correct analysis &amp;quot;modulator comprising ... means ...&amp;quot; or &amp;quot;transistors each having ... means ...&amp;quot;? A path through the network of Figure 3 shows that a modulator can have a direct voltage source but no such path exists for transistors. The appropriate instantiation of the network is created and the unacceptable syntactic path is eliminated from further consideration.</Paragraph>
      <Paragraph position="4"> Processing is complete when the text has been consumed and all the concepts from the text are connected to the topic of patent, though ungrammaticality may cause an earlier end.</Paragraph>
    </Section>
  </Section>
  <Section position="6" start_page="0" end_page="0" type="metho">
    <SectionTitle>
5.0 CONCLUSION
</SectionTitle>
    <Paragraph position="0"> Our understanding of language and cognitive processes is growing and novel programming languages are developing. With this knowledge and these tools, we are getting closer to viable natural language systems in limited domains.</Paragraph>
    <Paragraph position="1"> There are other developments that will contribute to building natural language systems, namely the decreasing cost and increasing power of hardware. Also advances in computer-aided design give promise of cost-effective special purpose machines with hardware routines for processes now implemented in software (Fahlman (1979)). More power will certainly aid in constructing language understanding systems. But the power will be wasted if it is used to attack a problem that can be resolved by some other approach, say by constructing an object-oriented system as presented above.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML