File Information
File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/93/h93-1006_metho.xml
Size: 17,156 bytes
Last Modified: 2025-10-06 14:13:19
<?xml version="1.0" standalone="yes"?> <Paper uid="H93-1006"> <Title>A PORTABLE APPROACH TO LAST RESORT PARSING AND INTERPRETATION</Title> <Section position="1" start_page="0" end_page="0" type="metho"> <SectionTitle> A PORTABLE APPROACH TO LAST RESORT PARSING AND INTERPRETATION </SectionTitle> <Paragraph position="0"/> </Section> <Section position="2" start_page="0" end_page="0" type="metho"> <SectionTitle> ABSTRACT </SectionTitle> <Paragraph position="0"> This paper describes an approach to robust processing which is domain-independent in its design, yet which can easily take advantage of domain-specific information. Robust processing is well-integrated into standard processing in this approach, requiring essentially only a single new BNF rule in the grammar. We describe the results of implementing this approach in two different domains.</Paragraph> </Section> <Section position="3" start_page="0" end_page="0" type="metho"> <SectionTitle> 1. Introduction </SectionTitle> <Paragraph position="0"> For best performance, natural language processing systems must be able to extract as much information as possible from their inputs, even inputs which cannot be fully processed. In order to do this, systems must be equipped with robust processing mechanisms. In addition, cases also occur in which the system has the ability to process an input, given sufficient time, but it is not desirable to allow unlimited amounts of processing time. In this paper we describe an approach to robust processing which is domain-independent in its general architecture, but which can be easily customized to particular domains by simply listing key words and/or key concepts. The approach uses the extensive grammar already available to the system for standard processing but augments it with a special BNF rule, called &quot;backup&quot;, which is able to prune the wordstream while it searches for key concepts. Backup can be triggered either by a failure of normal parsing or by timing out. This approach has been implemented in two distinct domains. In one of these domains, when sufficient time is allotted to attain maximal performance, backup results in an 18% improvement in score. We describe the general approach, discuss how differences in the data in each domain lead to slightly different implementations, and discuss our results.</Paragraph> </Section> <Section position="4" start_page="0" end_page="31" type="metho"> <SectionTitle> 2. Approach </SectionTitle> <Paragraph position="0"> The approach to robust processing which is described in this paper is implemented in the PUNDIT natural language processing system developed at Paramax Systems Corporation \[6, 1\]. PUNDIT includes a domainindependent, top-down parser \[7\] which is the primary component involved in robust processing. The key feature of robust processing in PUNDIT is that the parser is allowed to skip over words when it is unable to find a parse using every word. Skipping is an appropriate strategy for the data in the two domains we are working with, because parsing failures tend to be due to extraneous material such as interpolated irrelevant comments and false starts. Another possible strategy, relaxation of constraints as suggested by \[19\], is less appropriate for the data we have examined, since few parsing failures are due to violation of grammatical constraints. Skipping over words has also been implemented in the robust parsing strategies of Seneff \[15\] and Strzalkowski \[18\]; our approach differs from these in that in addition to skipping, it provides a simple way of taking domain-specific knowledge into account in the skipping process. That is, when an analysis is not possible using every word, the system begins searching through the wordstream for keywords (or words denoting key concepts), which are simply listed in a file. The use of keywords permits the system to make use of the domain-specific knowledge that certain words or concepts are important in the domain. In fact, in a mature domain, the list of keywords and concepts can be automatically generated from the system's semantic interpretation rules.</Paragraph> <Paragraph position="1"> Because the backup mechanism is implemented by adding a single new BNF rule into the normal grammar, robust processing has been implemented in PUNDIT without losing the advantages of the broad-coverage syntactic grammar already in the system. This is in contrast to approaches like the template matcher discussed in \[8\] or the frame combiner discussed in \[16\] which are completely separate mechanisms from the standard linguistic processing components.</Paragraph> <Paragraph position="2"> In addition to inputs for which the system cannot find a parse using the standard algorithm, there are also cases where a complete analysis would be too costly in terms of time. The system can also invoke backup in these cases, using a variation of the timeout mechanism described in \[17\]. The timeout mechanism in \[17\] allocates an absolute amount of time per sentence; in contrast, PUNDIT'S timeout allocates time as a function of the number of words in the input sentence so as not to penalize relatively longer sentences.</Paragraph> <Paragraph position="3"> Previous approaches to robust processing have typically either focused solely on data from one domain \[8, 16, 15, 4\] or have implemented a domain-independent approach \[17\]. Both of these alternatives have disadvantages. Approaches which have been tested on only a single domain cannot be guaranteed to be extensible to other domains. Entirely new approaches may be required when the system is ported to another domain. On the other hand, the performance of domain-independent approaches may suffer in domain-specific applications because they are not able to use domain-specific knowledge to constrain the processing. Our approach differs from previous approaches in that, while the basic architecture is domain-independent, the approach also allows domain-specific knowledge to assist in the processing.</Paragraph> <Paragraph position="4"> We demonstrate the general applicability of the architecture by describing implementations in two distinct domains. Although the basic mechanism is the same in each domain, we also discuss differences in the implementation which follow from basic differences in the kind of data which must be processed.</Paragraph> </Section> <Section position="5" start_page="31" end_page="31" type="metho"> <SectionTitle> 3. Domains </SectionTitle> <Paragraph position="0"> We now briefly describe our two application domains, with emphasis on those properties of the domains which affect the details of implementing backup &quot;last resort&quot; processing.</Paragraph> <Section position="1" start_page="31" end_page="31" type="sub_section"> <SectionTitle> 3.1. Air Traffic Control </SectionTitle> <Paragraph position="0"> Air traffic control (ATC) involves oral communication, as controllers interact with pilots via radio, issuing commands which govern the movements of planes both on the ground and in the air \[3\]. Since the controllers are already speaking into microphones, their half of this dialogue is easy to capture in a high-quality signal. If this input can be understood, possible applications will range from intelligent indexing for archival purposes to real-time monitoring for safety and planning purposes.</Paragraph> <Paragraph position="1"> Utterances in the ATC domain tend to be short sequences of relatively independent commands. The range of possible commands is well-bounded, and controllers are trained to avoid expressing these commands in different phrasings. As a consequence, it is possible to separate utterances into their constituent commands with high reliability, and similarly, to resume processing at the next command if processing of the present command fails for any reason. Also, some commands may be irrelevant for a given application. For example, wind advisories could be ignored by an application only concerned with ground operations.</Paragraph> <Paragraph position="2"> A sample well-formed utterance follows: Delta seven forty six turn right heading two seven zero cleared to land runway two nine left.</Paragraph> </Section> <Section position="2" start_page="31" end_page="31" type="sub_section"> <SectionTitle> 3.2. Air Travel Information System </SectionTitle> <Paragraph position="0"> Our second domain is called ATIS (Air Travel Information System) \[12, 13, 11\]. This is basically a database query application. The input utterances are retrieval requests addressed to a database of information about flight schedules, fares, etc. This application has been set up by DARPA as an infrastructure for research in spoken language understanding.</Paragraph> <Paragraph position="1"> DARPA has arranged for the collection of data in this domain \[5\]. This data is spontaneous speech from naive users, who have no idea what phrasings will work and which will not. Thus, they use an extremely wide set of variations for each request, so that the system is expected to process inputs ranging from a vanilla Show me flights from Boston to Denver to I am going to have to go to Denver; I will be leaving from Boston, etc. Disfluencies are more prevalent in this domain, since the speakers are not trained users. Another feature distinguishing ATIS from ATC is that ATIS utterances, no matter how discursive they appear, normally constitute a single request. Therefore parse fragments created by the backup mechanism seldom correspond to individual commands as they do in the ATC domain; instead, a single request may give rise to several fragments which must be integrated during semantic and pragmatic processing 1.</Paragraph> <Paragraph position="2"> In both domains, since the input is spoken, there is the additional possibility of errors introduced by the speech recognition component. While the techniques discussed in this paper have obvious applicability to recovery from such errors, in what follows we will assume perfection on the part of the recognizer, and that all errors and disfluencies originate with the speaker. Note, however, that current recognizers do not include punctuation in their output, either within sentences or at the end of them. We-therefore have included no punctuation in our data.</Paragraph> </Section> </Section> <Section position="6" start_page="31" end_page="33" type="metho"> <SectionTitle> 4. Implementation </SectionTitle> <Paragraph position="0"> Grammars used with PUNDIT have at the top level a BNF rule for the &quot;center&quot; node. This rule is always a disjunction of possibilities; for example, in a toy grammar, the center rule might expand to either assertion or question.</Paragraph> <Paragraph position="1"> In typical application domains this rule is more complex, including perhaps compounds and/or fragments.</Paragraph> <Paragraph position="2"> One important fact about the disjuncts for the present discussion is that they are required to consume the whole input word string in order to succeed.</Paragraph> <Paragraph position="3"> In any grammar, our approach to robust parsing is implemented by adding one additional disjunct at the end of the center rule. We call this disjunct &quot;backup&quot;. The BNF rule for backup has the following form: * If positioned at a keyword, reset the time allotment if necessary, then retry the other center options, relaxing the requirement to consume the entire word string. If a parse is found, call the center rule on the remainder of the word string.</Paragraph> <Paragraph position="4"> * If not positioned at a keyword, or if a parse is not found in the previous step, skip to the next keyword if any, reset the time allotment if necessary, and call the center rule on the word string starting with the keyword. If no keyword is found, fail.</Paragraph> <Paragraph position="5"> The backup rule is entered either if normal parsing fails (i.e., none of the other disjunets of the center rule produce a parse consuming the whole word string), or if timeout occurs. Users specify an amount of time in the form of a number (possibly fractional) of seconds per word, so that longer inputs are given more time. Once time has expired, no rule will execute except the backup rule, which will reallot time based on the length of the remaining word string, and then proceed as described above.</Paragraph> <Paragraph position="6"> The opportunity for introducing domain knowledge to influence the behavior of the backup rule comes in the specification of the keywords. To discuss what we have done in the two domains we experimented with, we first need to introduce the PUNDIT knowledge base. This is simply a mapping of word tokens to a hierarchical set of concepts \[10\]. Synonyms usually denote the same concept. The &quot;is-a&quot; relation is defined over the hierarchy, so that a concorde is-a jet is-a plane, a propeller_plane also is-a plane, etc.</Paragraph> <Paragraph position="7"> The keywords used by backup can be specified as word tokens or as concepts. In the latter case, the concept is taken to refer to any word token that maps to the concept or any descendant of the concept in the knowledge base. Keywords may also be specified by syntactic category, e.g., determiners or tensed verbs may function as keywords.</Paragraph> <Section position="1" start_page="32" end_page="32" type="sub_section"> <SectionTitle> 4.1. Air Traffic Control </SectionTitle> <Paragraph position="0"> In the ATC domain, we designated only word tokens as keywords. Furthermore, the list of keywords was chosen manually with great care, and is not very extensive. The choices were dictated by tile semantics of the possible commands which controllers may issue, and the normal phraseology (defined by the FAA) for expressing those commands. The intent, which we were able to achieve to a large degree, was to have skipping to the next key-word be equivalent to skipping to the start of the next command. Most of the keywords are verbs, corresponding to the imperative form most often used to express commands.</Paragraph> </Section> <Section position="2" start_page="32" end_page="33" type="sub_section"> <SectionTitle> 4.2. Air Travel Information System </SectionTitle> <Paragraph position="0"> In contrast, the list of keywords for the ATIS domain is much larger, and consists mostly of concepts, which in effect makes it even larger in terms of words. The basic idea is not to skip over any word which might be useful. Thus we included prepositions, wh-introducers, and such word tokens, plus all the concepts known to the PUNDIT semantic interpreter for that domain. This list of concepts was obtained mechanically from the files driving the interpreter, followed by the removal of concepts which were descendants of other concepts in the list, for these would be redundant for the purposes of the backup procedure. As a consequence, the only words skipped are meaningless (to the semantic interpreter), including unknown words.</Paragraph> <Paragraph position="1"> An ATIS utterance normally constitutes a single database retrieval request. Therefore an additional step in this domain is to integrate the parse fragments obtained by the robust parsing procedure. We delegate this responsibility to the semantic and pragmatic interpreter \[14, 2\]. For those fragments which are complete sentences, no extensions are necessary. The interpreter merely treats them as distinct sentences coming in sequentially in the context of the ongoing dialogue.</Paragraph> <Paragraph position="2"> For true fragments we did need to add some new capability. We assume that the overall content of the utterance is either a request for some flights or some fares. For noun phrase fragments, either the head is a flight or a fare, or it is not. If it is, our normal reference resolution capabilities are sufficient to resolve the flight or fare with any other flight or fare in the context 2. If the head is not a flight or fare, flight and fare entities are explicitly generated into the context space maintained by the semantic interpreter, and the fragment is interpreted as a modifier of either the flight or the fare. Then normal reference resolution takes over. For example, the fragment afternoon ends up with the same semantic representation as does afternoon flight, and the system proceeds as before.</Paragraph> <Paragraph position="3"> 2This is because dialogues often proceed like the following: Show me flights from Boston to Denver. \[answer\] Show me just afternoon flights. So in effect, afternoon flights is treated as show afternoon flights \[2, 13, 11\].</Paragraph> <Paragraph position="4"> Prepositional phrase fragments are treated in a manner completely analogous to noun phrase fragments whose heads are not flights or fares. For example, in the afternoon becomes flight in the afternoon, and the system proceeds as before.</Paragraph> <Paragraph position="5"> The data for this domain has not warranted treatment of any other fragment types.</Paragraph> </Section> </Section> class="xml-element"></Paper>