File Information
File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/06/w06-1806_metho.xml
Size: 24,997 bytes
Last Modified: 2025-10-06 14:10:45
<?xml version="1.0" standalone="yes"?> <Paper uid="W06-1806"> <Title>Natural Language Understanding using Temporal Action Logic</Title> <Section position="3" start_page="0" end_page="0" type="metho"> <SectionTitle> 2 Temporal Action Logic </SectionTitle> <Paragraph position="0"> The Temporal Action Logic (TAL) is a non-monotonic temporal logic developed specifically for reasoning about actions and dynamical domains. The logic has its origin in the Features and Fluents framework developed by Sandewall (1994) but was given a new characterization in terms of circumscription by Doherty (1994).</Paragraph> <Paragraph position="1"> Many extensions since have turned TAL into a very expressive language capable of representing, among other things, actions with durations, context-dependent and non-deterministic actions, concurrency, and action side-effects. It also provides solutions to the frame, ramification and qualification problems. For a more detailed introduction to TAL the reader is referred to (Doherty et al., 1998).</Paragraph> <Section position="1" start_page="0" end_page="0" type="sub_section"> <SectionTitle> 2.1 TAL Narratives </SectionTitle> <Paragraph position="0"> Domains are described in TAL using fluents that represent properties of the world that change over time. World laws governing the evolution of fluents are expressed in narratives, which are high-level descriptions of observations, action schemas, and action occurrences. Narratives receive a semantics through a translation to the base language, which is an order-sorted classical first-order logic together with a circumscription policy described below. Figure 1 shows the well-known Yale shooting scenario expressed as a TAL narrative. A persistence statement (labelled per1) constrains the fluents alive and loaded's values to persist unless they are affected by some action. Action schemas (acs1 and acs2) use the reassignment operator R to make loaded true after performing the Load action and to make both loaded and alive false after performing the Fire action, but only if loaded was true when the action was initiated. An observation statement (obs1) initializes loaded and alive to false and true respectively. Finally, the two occurrence statements (occ1 and occ2) describe a world history in which the Load action is performed between time points 1 and 2, and the Fire action is performed between 3 and 4.</Paragraph> </Section> <Section position="2" start_page="0" end_page="0" type="sub_section"> <SectionTitle> 2.2 Occlusion </SectionTitle> <Paragraph position="0"> The key to the solution of the frame problem in TAL lies in the use of occlusion. When narratives are translated into classical logic an Occlude predicate is introduced and constrained to be true at those time points where fluents are forced to change their values due to reassignments. An additional no-change axiom is added that rules out</Paragraph> <Paragraph position="2"> any fluent changes at any time points when the fluent was not occluded. The final step is the minimization of the Occlude predicate and the action occurrences, expressing the default assumption that no unexpected fluent changes or spurious actions occur unless explicitly specified. The minimization is accomplished through the circumscription of those parts of the translated theory that contain action schemas and action occurrences respectively. null</Paragraph> </Section> </Section> <Section position="4" start_page="0" end_page="0" type="metho"> <SectionTitle> 3 System Architecture </SectionTitle> <Paragraph position="0"> NL1 carries on an interactive natural language text dialogue with the user, executing commands and answering queries about a simulated blocksworld.</Paragraph> <Paragraph position="1"> An important emphasis of the system architecture, depicted in Figure 2, is the use of declarative knowledge structures in the hope of building a system that is both convenient to adapt to new usage scenarios and that has a great ultimate potential for extension while at the same time retaining the basic implementation components.</Paragraph> <Section position="1" start_page="0" end_page="0" type="sub_section"> <SectionTitle> 3.1 Natural Language Module </SectionTitle> <Paragraph position="0"> The natural language module is responsible for the parsing of input sentences and the construction of a quasi-logical form. User interaction consists of natural language input and output through a minimal user interface in the form of a text top-loop.</Paragraph> <Paragraph position="1"> Sentence are read from a prompt and passed as text strings to the Linguistic Knowledge Builder (LKB) component. The LKB chart parser uses a Head-driven Phrase Structure Grammar (HPSG), based on the grammar in (Sag et al., 2003), to parse the input text string and build a feature structure representation. The grammar includes semantical relations, and the semantical part of the feature structure representation constitutes a Minimal Recursion Semantics (MRS) structure.</Paragraph> <Paragraph position="2"> One of the benefits of this set-up is that natural language generation can be achieved by running the chart parser &quot;in reverse&quot; using the same grammar that was used when parsing. The task of generating a response is then reduced to the task of constructing a suitable MRS structure representing an answer to the user's request.</Paragraph> <Paragraph position="3"> The HPSG grammar forms a declarative knowledge source that is easily adapted to new vocabularies, by changing the lexicon of words and semantical relations, and that has a great potential for extension, e.g., moving towards broad coverage as in the English Resource Grammar project (Copestake and Flickinger, 2000).</Paragraph> </Section> <Section position="2" start_page="0" end_page="0" type="sub_section"> <SectionTitle> 3.2 Knowledge Representation Module </SectionTitle> <Paragraph position="0"> The knowledge representation module holds a representation of both the basic world laws and the particulars of the current scenario, but also a history of the developments of the world during the ongoing dialogue.</Paragraph> <Paragraph position="1"> The world model, together with action specifications and an initial state is encoded as TAL formulas and stored in the blocksworld knowledge base. A continually evolving TAL narrative of action occurrences represents the world history and is stored in the state knowledge base. Any sentence that enters the system reaches the knowledge representation module in the form of an MRS structure that needs to be transformed into a full logical form before it is used in reasoning. The MRS structure is passed to the translation component that implements a partial translation by performing a number of transformations, as described in Section 4, that result in a TAL formula. If the natural language module identified the sentence as an imperative command, the TAL formula is appended to the narrative, but if the sentence was identified as a proposition or a question, a reasoning problem is formed by combining the corresponding TAL formula with the background blocksworld theory.</Paragraph> <Paragraph position="2"> By using TAL as our knowledge representation language we are able to express a wide range of common-sense reasoning scenarios and ensure that the system is not limited to the relatively simple blocksworld, even though it is used as an illustrative example.</Paragraph> </Section> <Section position="3" start_page="0" end_page="0" type="sub_section"> <SectionTitle> 3.3 Automated Reasoning Module </SectionTitle> <Paragraph position="0"> Performing automated reasoning in the Temporal Action Logic is not trivial since it is a non-monotonic logic with a semantics based on circumscription. When Sandewall developed the basis of the formalism he was concerned more with assessing its correctness rather than performing automated reasoning. Later Doherty and Lukaszewicz (1994) showed how the semantics could be expressed using circumscription and, in (Doherty, 1994), that under certain restrictions the circumscription policy could be reduced to predicate completion (defined by Lifschitz (1994)) which is computable and results in an equivalent first-order characterization.</Paragraph> <Paragraph position="1"> These methods open up the possibility of reasoning in TAL using a very flexible compilation approach that fits in nicely with the TAL high-level narrative description to base language translation. The idea is to choose an existing domain-independent automated reasoning paradigm and develop a compilation step from the TAL base language into a language compatible with that paradigm. The most obvious example would be a compilation into regular first-order logic and the application of resolution theorem provers, but other options are certainly possible as will be described below. The flexibility lies in the postponing of the expressibility vs. tractability tradeoff decision to the time at which some specific application domain is fixed, or even later, to the time at which a specific reasoning problem needs to be addressed.</Paragraph> <Paragraph position="2"> For example, one would like to detect if the theory is Horn and then use a logic program compilation while still retaining the possibility to automatically fall back to first-order or even higher-order logic if necessary. Such a classification of the complexity of the reasoning problem is not possible in general, but even an incomplete version might be tremendously useful in certain domains.</Paragraph> <Paragraph position="3"> Our implementation includes two different compilation mechanisms. One results in a first-order theory that is further encapsulated (adding explicit theorem proving strategies and other configurations) to form an input problem to SNARK (Stickel, 2005), a resolution theorem prover with support for, among other things, sorts and answer (1) B1 is on the table.</Paragraph> <Paragraph position="4"> Ok.</Paragraph> <Paragraph position="5"> (2) The table supports B1.</Paragraph> <Paragraph position="6"> Yes.</Paragraph> <Paragraph position="7"> (3) Does the table support B2? I don't know.</Paragraph> <Paragraph position="8"> (4) B2 is on the table and B3 is on B1.</Paragraph> <Paragraph position="9"> Ok.</Paragraph> <Paragraph position="10"> (5) Does the table support B3? No.</Paragraph> <Paragraph position="11"> (6) What blocks are clear? B3 is clear and B2 is clear.</Paragraph> <Paragraph position="12"> (7) Pick up B3.</Paragraph> <Paragraph position="13"> Ok.</Paragraph> <Paragraph position="14"> (8) Put B3 on the table.</Paragraph> <Paragraph position="15"> Ok.</Paragraph> <Paragraph position="16"> (9) What blocks are on the table? B2 is on the table and B1 is on the table extraction. The other compilation results in a logic program that is further specialized into a program for Allegro Prolog2, a Prolog environment that is fully integrated in the Allegro Common-Lisp development environment. In both cases, depending on the question type, the result is the deduction of true or false, or the extraction of answer terms from a proof of the query formula. These answer formulas are passed back to the natural language module where they are used in the construction of an answer MRS that generates a natural language response, closing the loop.</Paragraph> </Section> <Section position="4" start_page="0" end_page="0" type="sub_section"> <SectionTitle> 3.4 Example Dialogue </SectionTitle> <Paragraph position="0"> The example dialogue in Figure 3 illustrates the working system and consists of numbered interactions between an operator and the software. The simplest sentence type is a proposition stating a fact about the world, as in sentence 1. NL1 parses the sentence, translates the result into a TAL formula, and adds it to the knowledge base. Since the knowledge is expressed using logic we can define new relations in terms of old relations using axioms, as in the support relation of sentence 2. Before accepting new knowledge NL1 uses its reasoning capabilities to determine if the new information contradicts the current knowledge or if it</Paragraph> <Paragraph position="2"> the table supports B1 is entailed by the previous fact and the axioms defining support, so the system simply replies yes. In sentence 3 we demonstrate the evaluation of a simple yes/no-question, although in this case the answer is not known since nothing has been said about block B2 and NL1 makes no closed world assumptions. Sentence 4 adds B2 to the table and B3 on top of B1 so that when the question in sentence 5 is evaluated it follows, from the fact that B3 is on B1 and the fact that a block cannot be in two places at the same time, that the table can not support B3. A much more complex type of questions is what-questions.</Paragraph> <Paragraph position="3"> Sentence 6 poses a what-question about another defined relation, clear. The system processes the question, as described in detail in Section 4, and constructs an answer MRS from which a complex natural language reply is generated. The sentences in 7 and 8 are recognized as imperative and are parsed into TAL action occurrences that are added to the world history. Finally, the question in sentence 9 demonstrates that the new blocksworld state conforms to the changes specified for the performed actions in the background action theory listed in Figure 7.</Paragraph> </Section> </Section> <Section position="5" start_page="0" end_page="0" type="metho"> <SectionTitle> 4 Sentence Processing </SectionTitle> <Paragraph position="0"> The sentence processing can be described by an (incomplete) algorithm, shown in figure 4, that translates a feature structure representation of an MRS into a TAL formula. To illustrate the algorithm let us assume that a user enters the question in Figure 5a. The LKB component parses the string and the MRS in Figure 5b is extracted from the resulting feature structure. The MRS Trans component uses the algorithm to interpret the sentence as a TAL formula, starting by con-</Paragraph> <Paragraph position="2"/> <Paragraph position="4"/> <Paragraph position="6"> verting the MRS structure to a relation list, shown in Figure 5c. Our simple treatment of quantifiers assigns the scope of the entire formula while ordering each quantifier as they appear in the relation list and treating the as an existential quantifier, as in Figure 5d. In Figure 5e, the TAL time operator has been applied to create a temporal formula replacing the event variable e1 with the designated constant now that is updated during the dialogue.</Paragraph> <Paragraph position="7"> Finally, to take advantage of the fact that TAL is an order-sorted logic, one-place relations that are identical to sorts defined in the current background theory are compiled into a symbol table and removed from the formula in Figure 5f.</Paragraph> <Paragraph position="8"> At this point we have arrived at the TAL formula representing the user's question. Let us further assume that we are using the TAL blocksworld formalization shown in Figure 7, and that the current state of the dialogue interaction has produced the narrative history represented by the two final occ statements in the figure. These components together form a reasoning problem and any answer to the original question is an instantiation of the variable(s) selected by the special answer predicate in the question formula that satisfies the blocksworld specification together with the current world history.</Paragraph> <Paragraph position="9"> Whatever method chosen, to solve the specific reasoning problem, will produce one or more answer formulas as show in Figure 6a. The answers are used to instantiate the question formula from Figure 5f to form the TAL representation in Figure 6b, representing the answers to the question.</Paragraph> <Paragraph position="10"> Again the MRS Trans module applies a number of processing steps, this time to construct an MRS structure representing the answers. First, the TAL time operator is translated into event variables and the quantifiers flattened to relations in Figure 6c.</Paragraph> <Paragraph position="11"> In Figure 6d, the references to the different blocks are made explicit in the relation list and the existential quantifier relations are translated totherelations, assuming the answer recipient knows what instance is meant. Finally, the MRS feature structure in Figure 6e is built and passed to the LKB module which uses the HPSG grammar to generate a natural language sentence corresponding to it. This sentence, displayed in Figure 6f, represents the natural language answer to the natural language question posed in Figure 5a.</Paragraph> </Section> <Section position="6" start_page="0" end_page="0" type="metho"> <SectionTitle> 5 Discussion </SectionTitle> <Paragraph position="0"> Natural language understanding was among the first research topics of artificial intelligence and has continued to be of great importance. As a consequence a large number of natural language understanding and dialogue systems have been built.</Paragraph> <Paragraph position="1"> Even though our system does not contribute any new techniques, it is a novel combination of existing techniques that form an environment in which a variety of linguistic and knowledge representation problems can be attacked.</Paragraph> <Paragraph position="2"> Both the HPSG grammar and the TAL background knowledge are declarative knowledge sources that can be updated, incrementally and iteratively extended, or adapted to other use scenarios, independent of each other and most of the implementation components. They are also very expressive, contributing to the generality of the approach and to the goal of achieving representational adequacy. Equally important to expressiveness are issues of efficiency. Through a flexible compilation scheme for automated reasoning the tradeoff between expressiveness and efficiency is not set in stone but can be adapted to the task at hand. Different compilation steps can be implemented from different subsets of TAL to different automated reasoners and the choice of which compilation step to use is postponed.</Paragraph> <Paragraph position="3"> While we have defended our design decisions we also acknowledge that they do give rise to some disadvantages. Relying on deep parsing might result in an instance of what is often called the &quot;brittleness&quot; problem where sentences slightly outside the competence of the grammar is totally incomprehensible and any grammatical error has similar results. We agree to some extent with this objection but note that there is interesting work on a Robust Minimal Recursion Semantics (Copestake, 2003) that would help integrate deep and shallow parsing. However, it should be pointed out that certain problems can never be solved using shallow methods and the mixing of deep and shallow methods can only obscure their solutions. Although generality was listed as an advantage, we also include it as a system drawback and possible point of attack by critics. A more general system will always be worse at any given task than a less general system that is more specialized toward that task. This rule weakens as the complexity of the task grows, but holds for most problems dealt with in current AI research.</Paragraph> <Section position="1" start_page="0" end_page="0" type="sub_section"> <SectionTitle> 5.1 Evaluation </SectionTitle> <Paragraph position="0"> Ultimately deciding on the viability of a framework such as NL1 should depend in some way on evaluation, but we have yet to perform any structured evaluation attempts and it is far from obvious how they would be implemented. The emphasis is not on the immediate improvement in the measurable efficiency of some real-world task like natural language translation, or the maximum coverage of a large number of question answering tasks, but instead on the investigation and exploration of a potentially fruitful marriage between natural language and knowledge representation technologies.</Paragraph> <Paragraph position="1"> At the same time, it is clear that there are obvious improvements that could be made to the current NL1 system, and that one would like to formulate some measure that made what is intuitively obvious, objectively evident. To this end we propose evaluations through a problem collection.</Paragraph> <Paragraph position="2"> Even a very simple system can answer an infinite number of questions correctly if all that is varied is some trivial question property such as the name of the blocks in a blocksworld. It is not sufficient to count the number of sentences that are answered without further classifying them 33 KRAQ06 according to some complexity dimension, which is easier said than done. But even without this classification, one carefully selected sentence, illustrative of some complexity of language or reasoning, can still make or break a system. We propose that one creates and maintains an on-line collection of such examples, similar to the common-sense reasoning problem page (Morgenstern, 2005), the logically reasoning agents problem page (Thielscher, 2005), or the logic modelling workshop (Sandewall, 2005). New examples, illustrating new difficulties, would be submitted by different researchers and added to the problem page (with moderation). Evaluating different versions of the same system would then simply be accomplished by noting that the new version solves some additional question (ignoring, among others, issues of efficiency). Comparisons between systems would be entirely possible if one system subsumes the sentences correctly handled by the other system and possible with subjective results if the systems had a partial overlap.</Paragraph> </Section> <Section position="2" start_page="0" end_page="0" type="sub_section"> <SectionTitle> 5.2 Limitations and Future Work </SectionTitle> <Paragraph position="0"> We think this project has only scratched the surface of what is possible to accomplish using these techniques and that the proposed architecture has great potential. This section will point out the most important current limitations as well as our plans for continued development.</Paragraph> <Paragraph position="1"> An obvious improvement is an extension of the HPSG grammar coverage to make the dialogue more varied and robust. The grammar does not at present even cover all the language constructs described in the book that it is based on (Sag et al., 2003) and there are certainly other HPSG grammar work that can be adapted to our system to further complement and extend coverage. An interesting experiment would be a coupling to the English resource grammar (Copestake and Flickinger, 2000).</Paragraph> <Paragraph position="2"> Another extension that would put our approach to the test is a method of dealing with different forms of reference. We envision that, in addition to the domain background theory, the knowledge representation module will contain a model of the ongoing dialogue expressed using the same temporal action logic. Reference resolution would then be the solution of additional reasoning problems where objects that simultaneously fulfilled declarative constraints from dialogue factors, such as recency, and background knowledge, would be retrieved.</Paragraph> <Paragraph position="3"> Time is a central concept in the system, yet at present this is not taken advantage of. Questions are all in present tense, even though everything is set up so as to support one talking to the system about past actions and time.</Paragraph> <Paragraph position="4"> The current implementation recognizes commands of action and executes them, but such commands are restricted to simple atomic operations and the system can not by itself plan a sequence of actions in response to a user request. As part of another project, we are working on extending TAL to incorporate composite actions and action sequences or plans. Such work would fit naturally in the framework we have described here and would enable a seamless transition between requesting simple actions and requesting complex actions, possibly requiring the use of deductive planning, without extending the system architecture with a special purpose planner.</Paragraph> <Paragraph position="5"> The compilation approach to automated TAL reasoning is inherently suitable for experimentation. While we have already performed work in this direction, developing several different compilations to first-order logic and another to logic programs, we do not expect to run out of ideas in this area. An especially interesting one is the use of deduction system alternatives to resolution, such as natural deduction, that might be more suitable for the kinds of inferences needed in the logicist approach to natural language understanding.</Paragraph> </Section> </Section> class="xml-element"></Paper>