File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/85/e85-1037_metho.xml

Size: 22,559 bytes

Last Modified: 2025-10-06 14:11:42

<?xml version="1.0" standalone="yes"?>
<Paper uid="E85-1037">
  <Title>A PROBLEM SOLVING APPROACH TO GENERATING TEXT FROM SYSTEMIC GRAMMARS*</Title>
  <Section position="3" start_page="0" end_page="0" type="metho">
    <SectionTitle>
THE FUNDAMENTAL RELATIONSHIP
</SectionTitle>
    <Paragraph position="0"> I. &amp;quot;The central nature of intelligent Problem solving is that a system must construct its solution selectively and efficiently from a space of aiternatlves.&amp;quot; \[Hayes-Roth et al., 1983) 2. &amp;quot;We shall define language as 'meaning potential': that is as sets of options or alternatives, in meaning, that are available to the speaker-hearer.&amp;quot; \[Halliday in deJoia et al., 1980, I~72) Compare these two quotations. Notice that both AI problem solving and systemic grammar have at their very core the idea of choosing from alternatives. Systemic grammar is probably unique in having such emphasis on the idea of choice; or in different terminology, systemic grammar is distinguished in allowing the paradigmatic mode of description to dominate over the syntagmatic \[see Halliday et al., 1981, p. 19). Thus, this is a special relationship between systemic grammar and AI problem solving.</Paragraph>
    <Paragraph position="1"> This fundamental relationship can be exploited because systemic grammar provides knowledge \[in the AI sense) about the various linguistic alternatives, and AI problem solving provides computational techniques for choosing between the alternatives, given the appropriate knowledge. The text generation approach described here is simply the standard AI knowledge-based problem solving methodology, with a systemic grammar acting as Dart of the knowledge base.</Paragraph>
  </Section>
  <Section position="4" start_page="0" end_page="252" type="metho">
    <SectionTitle>
KNOWLEDGE REPRESENTATION
</SectionTitle>
    <Paragraph position="0"> One of the manifestations of this fundamental relationship between AI problem solving and systemic grammar is a common representation of knowledge. Both of these fields represent the interdependencies between the various alternatives as &amp;quot;condltion/effect&amp;quot; relationships.</Paragraph>
    <Paragraph position="1">  problem solving The last decade has produced problem solving techniques which use domain-specific knowledge to guide the problem solving process. Problem solving knowledge is often expressed as condition/effect rules. For instance, a medical problem solver may have the rule: if then a patient has symptoms X, and Y drug A should be administered.</Paragraph>
    <Paragraph position="2"> Here if the conditions Ithe symptomsJ are satisfied, the problem solver can infer that drug A should be given. At this point other rules may be involved: if a drug should be administered and not in stock then...</Paragraph>
    <Paragraph position="3"> it is The problem solver is forming a chain of inferences which leads toward the solution. This is called &amp;quot;forward chaining&amp;quot;.</Paragraph>
    <Paragraph position="4"> Condition/effect rules can also be used to reason from the effects back to the conditions. SUDDOSe we have a rule: if then a surface is hot and compound B is applied the surface will be made Permanently non-reflective.</Paragraph>
    <Paragraph position="5"> If a problem solver has a goal to make a surface non-reflectlve, it can see from the effects that this rule will achieve the goal. The conditions of the rule are set as subgoals, and the problem solver will try to find rules to achieve these. Rules must then be found that have the effects of heating the surface and applying the compound. Here the problem solver is working backward from the solution. This is called &amp;quot;goal-directed backward chaining&amp;quot;.</Paragraph>
    <Paragraph position="6"> s~stemic grammar Systemic linguistics was developed in the early sixties by M.A.K. Halliday, although its roots in sociology and anthropology extend back much further. The emphasis of systemic linguistics has not been on the structure of language, but on its function; systemicists are not so much interested in what language 'looks llke', as in how it is used. They are interested in how language is used to ask questions and make statements, how language can be used to relate 'who did what to whom', and how language ties itself to previous discourse.</Paragraph>
    <Paragraph position="7"> The relationship between this functional view of language and the structural tradition is analogous to the relationshi~ between Physiology and anatomy*, and is equally complementary. This functional perspective has led to a different conceptualization of what language is, and how it should be described.</Paragraph>
    <Paragraph position="8"> The most important knowledge structure in systemic grammar is the 'system' ~ this is where the theory gets its name. A system is simply a mutually exclusive choice between a set of alternative features. Figure I shows a system that represents a choice between a marked- and unmarked-wh-theme.  unmarked-wh-theme wh- I IWh / Tdegpicall o i I I~ &amp;quot; Flnltel' i marked-wh-theme I-r77-.. 1  combination of features that must be chosen before the particular choice is appropriate. In this case the entry condition is simply the feature wh-. So the clause must be a wh- clause before the choice between a marked- or unmarked-wh-theme is relevant. The boxes contain what are called 'realization rules'. These specify the syntactic consequences of choosing the associated feature. &amp;quot;Wh / Topical&amp;quot; is read: &amp;quot;the Wh element is conflated with the Topical&amp;quot;, meaning that the Wh and Topical are realized by the same item. &amp;quot;Wh &amp;quot; Finite&amp;quot; is read: &amp;quot;the Wh element is adjacent to the Finite element&amp;quot;, meaning that the Wh element immediately precedes the Finite element in the clause.</Paragraph>
    <Paragraph position="9"> As well as systems, systemic grammars may contain what Mann \[19831 calls &amp;quot;gates'. A gate also has some logical combination of features acting as entry conditions.</Paragraph>
    <Paragraph position="10">  In Figure 2 the curly bracket means AND, and the square bracket means OR. A gate also may have realization rules. Here the Finite element is constrained to be some form of 'does': &amp;quot;does&amp;quot;, &amp;quot;does not&amp;quot; or &amp;quot;doesn't&amp;quot;. The significant difference  between systems and gates is that gates do not involve a choice.</Paragraph>
    <Paragraph position="11"> * This analogy was probably first made by Firth (1957) and has been used several times since -- see \[Winograd, 1983, p.287J  Now consider these two constructs from a problem solving point of view. A feature that is part of a system can be &amp;quot; interpreted as a condltion/effect rule. The conditions are simply the entry conditions of the system; the effects are choosing the feature, and doing whatever the realization rules say. This means that these features can be interpreted as problem solving rules and put at the disposal of the problem solver. Again it must be stressed that a system involves choice. From a problem solving point of view choices should be avoided whenever possible, in case the wrong choice is made. Notice if a system feature is used for backward chaining the choice is not explicitly considered. Suppose there is a goal to choose unmarked-wh-theme. Since the problem solver can interpret the system features as condition/effect rules, it sees that there is a rule called unmarked-wh-theme that achieves this goal as one of its effects. The problem solver begins to backward chain by invoking this rule and setting its condition, wh-, as a subgoal. The feature marked-whtheme was never explicitly considered. Similarly, features that are gates can be interpreted as forward chaining condition/effect rules. In Figure 2, if the entry conditions are satisfied, the does rule fires, choosing does and constraining the Finite element.</Paragraph>
  </Section>
  <Section position="5" start_page="252" end_page="254" type="metho">
    <SectionTitle>
THE METHOD
</SectionTitle>
    <Paragraph position="0"> The last section showed that features from systemic grammars can be interpreted as a condition/effect rule of the type used by AI Problem solvers, regardless of whether they are part of a system or a gate. An AI problem solver can thus use a systemic grammar as part of its knowledge base, and solve grammatical problems in exactly the same way as it solves medical problems using medical knowledge, or chemistry problems using chemistry knowledge.</Paragraph>
    <Paragraph position="1"> an examDle Figure 3 is a simplified excerpt from a systemic grammar. Suppose, for the moment, that the semantics wants to choose unmarked-declarative-theme and operative. The grammar provides rules that achieve these goals as Dart of their effects. The feature unmarked-declarative-theme can be thought of as a rule that chooses that feature and conflates the Subject with the Theme. This rule has, however, the condition declarative. This is set as a subgoal which can be achieved by another rule tl~at in turn has the condition indicative. In this way the problem solver backward chalnsfrom unmarked-declaratlve-theme through declarative, through indicative, through finite, to clause. At this point the backward chaining stops because clause has no conditions. The problem solver also backward chains from operative through effective to clause. Once clause is chosen, the gate theme fires \[the only instance of forward chaining in this example).</Paragraph>
    <Paragraph position="2"> Every time a rule is used the 'realization rules' in the effects are accumulated, gradually constraining the structure of the clause. In the example, the Agent has been constrained to be the leftmost constituent in the clause. The semantics will choose other features of course, from parts of the grammar not shown here, and after further forward and backward chaining, the clause will be completely determined.</Paragraph>
    <Paragraph position="3">  The careful reader may have noticed that it is possible for the semantics to start the same process with the goal &amp;quot;move the agent into the theme Position&amp;quot; \[conflate Agent and ThemeJ, assuming there is a rule expressing the transitivity of conflation. The transitivity rule would set as subgoais: &amp;quot;conflate Agent with X&amp;quot; and &amp;quot;conflate Theme with X&amp;quot;, where X could be instantiated to Subject. From there the problem solving would proceed as before. However, this would require far too much inference for such a simple goal. First, the transitivity would have to be worked out correctly. Second, there are likely to be other rules with the same realization rules, but which would lead to conflicts, and backtracking.</Paragraph>
    <Paragraph position="4"> In problem solving, if a simple goal requires too much inference, its solution can be 'compiled' \[Brachman, 1983J. Here, the semantics may have a rule that says: if then there is a goal to make a statement and a goal to move the agent into the theme Position choose unmarked-declarative-theme and operative.</Paragraph>
    <Paragraph position="5"> This use of compiled knowledge to actually choose features from the grammar corresponds to the systemic idea of 'preselection'. Preselection is an important part of systemic theory, being the vehicle of realization across network boundaries. Systemic grammar:adopts ... the general perspective on the linguistic system you find in Hjelmslev, in the Prague school, with Firth in the London school, with Lamb and to a certain extent with Pike - language as a basically tristratai system: semantics, grammar, phonology. \[Halliday, 1978, P.39J Each level must Pass down information to the level below. Realization rules at the higher level Dreselect features from the next level below. The semantic stratum \[described in the next sectionJ preselects features from the grammatical stratum \[e.g. unmarked-declarative-theme and operative in the example aboveJ. Simliarly, the grammatlcai stratum preselects phonologlcal/graphologlcal features.</Paragraph>
    <Paragraph position="6"> Preselection is also used to interface the different ranks at the grammatical level \[clause, group and wordj. The colon in Figure 2 is the symbol for preseleetlon. Thus the feature does at the clause rank preselects the feature does from the auxiliary network at the word rank. If, for instance, the features reduced and negative are also preseleoted, the Finite element will be realized as &amp;quot;doesn't&amp;quot;.</Paragraph>
    <Paragraph position="7"> Returning to Figure 3, compare this backward chaining approach to Mann's \[1983) NIGEL system. NIGEL begins at the left hand side of the network and works its way towards the right. It starts by choosing the feature clause. Then it sees that it must choose between finite and non-finite. There is a semantic 'choice-expert' associated with this system which cannot make the choice without specific information about the context and the communicative goals, The choice expert gains this information by passing messages to the 'environment'. In this case the answer returned from the environment will indicate that finite should be chosen. Another choice expert will now choose between indicative and imperative and so on.</Paragraph>
    <Paragraph position="8"> Whether or not this is a valid or interesting way to do text generation is not at issue here.</Paragraph>
    <Paragraph position="9"> From a computational point of view NIGEL has some drawbacks. Most importantly, an explicit choice must be made for every system encountered during the process. For large grammars, this will number in the hundreds, and will result in a large overhead. In contrast, the preselection - backward chaining approach outlined in this paper greatly reduces the number of explicit choices, The reason these choices are avoided here is that the problem solving process is ~oal-directed. The semantic stratum chooses some features from the right hand side of the network, which greatly reduces the number of Possible paths through the network from the very start.</Paragraph>
    <Paragraph position="10"> It could be argued that this kind of goal-directed search is non-deterministlc because systems may have disjunctive entry conditions, There is, however, an AI problem solving technique which has been developed for this purpose: least commitment \[Stefik et al., 1983~. Least commitment is simply the principle of not making any choices until absolutely necessary. Whenever a disjunctive entry condition is encountered, a decision must be made about which subgoal to set. There is no requirement that the decision be made at that particular instant, so it is suspended until one of the subgoals is set as part of another chain in inference \[gratuitously solving the original problemJ. Of course there will be cases where none of the subgoals \[entry conditions) are part of another inference. In these cases, it must be assumed that the semantics will preselect a feature corresponding to one of the subgoals.</Paragraph>
    <Paragraph position="11"> Clearly this whole text generation method relies on the semantic level to preselect the appropriate grammatical features. The next section will briefly look at this semantic level.</Paragraph>
  </Section>
  <Section position="6" start_page="254" end_page="254" type="metho">
    <SectionTitle>
SEMANTICS
</SectionTitle>
    <Paragraph position="0"> No motivation for the stratified approach adopted by systemic grammar will be given here, except pointing out that the role of the semantic stratum is to interface the extra-linguistic with the grammatical \[Halliday, 1978). In order to preselect the correct features from the grammar, this level must contain a considerable amount of knowledge \[in the AI sense) relating grammatical features to extra-lingulstic factors.</Paragraph>
    <Paragraph position="1"> In this section we will look at one particular organization of the semantic stratum, as presented in \[Halliday, 1978). Halliday organized his semantic stratum around the idea of 'register': It refers to the fact that the language we speak or write varies according to the type of situation ... What the theory of register does isattempt to uncover the general principles which govern this variation, so that we can begin to understand what situational factors determine what linguistic features. \[Halliday in deJoia st al., 1980, #764) Halliday uses the same system network notation to describe the semantics. Figure 4 \[adapted from \[Halliday, 1978)) describes the control strategies that a mother can use on her child.</Paragraph>
    <Paragraph position="2"> The features of a semantic system network, llke those of the grammatical networks, have realization rules ~ including preselection. For instance the semantic feature re4ection Dreselects the features which will make the hearer the Medium \[Affected), and realize it with the pronoun 'you' \[by preselecting from the nominal group and noun networks). The semantic feature decision preselects, for instance, the clause feature declarative. The semantic feature resolution Preselect3 the features present-in and present to give this type of threat its tense construction -e.g. &amp;quot;you're going upstairs&amp;quot;, &amp;quot;I'm taking you upstairs&amp;quot;. Similarly, obligation preselects necessary passive modulation \[Halliday, 1970) -- e.g. &amp;quot;I'll have to take you upstairs&amp;quot;, &amp;quot;you'll have to go upstairs&amp;quot; \[Halliday, 1978).</Paragraph>
    <Paragraph position="3"> Unfortunately, very little work has been done in the area of register, even by Halliday and his colleagues, so no large portions of a semantic stratum have been built. However, this example illustrates the idea.</Paragraph>
  </Section>
  <Section position="7" start_page="254" end_page="254" type="metho">
    <SectionTitle>
ADVANTAGES
</SectionTitle>
    <Paragraph position="0"> The backward chaining approach outlined here has several advantages. First, this method does not involve any linguistic sacrifices, since an established linguistic formalism is utilized. Systemic grammar was developed by linguists for linguistic purposes, and is used here in a totally unadulterated form. Nothing llnguisticaily ad hoc has been introduced for computational reasons.</Paragraph>
    <Paragraph position="1"> Second, no computational sacrifices have been made to accommodate the linguistic formalism.</Paragraph>
    <Paragraph position="2"> State-of-the-art computational techniques are being exploited at all stages of the problem solving process. null Third, the approach is parsimonious. There is no need for a sPecial-purpose text generation component. Other methods involve an AI problem solver that does the extra-linguistic work and perhaps the high-level 'text-plannlng', then passes a specification off to a special-purpose mechanism that processes the grammar. Here the AI problem solver can directly process the grammar; eliminating the special purpose component, and avoiding the kind of message passing that NIGEL, for example, must do.</Paragraph>
  </Section>
  <Section position="8" start_page="254" end_page="255" type="metho">
    <SectionTitle>
PROJECT STATUS
</SectionTitle>
    <Paragraph position="0"> At present, this approach to text generation is being tested on a large systemic grammar. The grammar has been collected from a variety of sources \[Mann/Halliday) \[Kress, 1976J \[Halliday &amp; Hasan, 1976) \[Winograd, 1983J, and contains about six hundred grammatical features. Fragments of grammar usually appear in the linguistic literature as 'system networks'. These are entered as LISP data structures, and translated by a three page LISP program into OPS5 production rules, lOPS5 is a widely used production system that was used to implement, for example, RI \[Gaschnig et al., 1983JJ.</Paragraph>
    <Paragraph position="1"> once the grammar is in the form of OPS5 rules, OPS5 can perform forward and backward chaining directly. The rest of the system consists mostly of OPS5 rules to act on the realization rules of the grammar, and to output the text as it is being generated. null The interface between the grammar and the  semantics has been implemented, namely preselection. Since preselectlon is done via realization rules, it is implemented by a small group of OPS5 rules as just mentioned.</Paragraph>
    <Paragraph position="2"> Although the interface between the grammar and the semantics has been implemented, the semantic stratum itself has not. This means that to test the approach, those features that would have been preselected by the semantics must be preselected by hand.</Paragraph>
    <Paragraph position="3"> Another limitation at the moment is that there is no graphological level. This means that the output does not contain punctuation, capitals, the word &amp;quot;an&amp;quot;, and so on.</Paragraph>
    <Paragraph position="4"> To put all this in perspective, recall that systemic linguistics stratifies language into the semantic, the grammatical, and the graphological \[or if working with speech, phonologicalJ strata. Currently only the middle stratum, the grammatical; has been implemented. Again it should be Pointed out that the interfacebetween the different strata \[preselectlon in each caseJ has been implemented as well.</Paragraph>
    <Paragraph position="5"> sample output Consider the context of a medical expert system that is trying to diagnose a patient's illness. Suppose there is a patient named Mary who has been having headaches and stiff neck muscles. The expert system hypothesizes that Mary has a fever, and tests this hypothesis by asking &amp;quot;Does Mary have a fever ?- At this point, the user, who we will assume is neither a medical or computing expert, can ask &amp;quot;WHY&amp;quot; \[did you ask me that question?J*. The test system at this stage can generate the fol&amp;quot; lowing response \[bars have been added to indicate clause boundaries).</Paragraph>
    <Paragraph position="6"> il well mary has been having headaches II on this basis perhaps she has a infection II this DOSSlbility would be SUPDorted by a fever II so we ask I does she have one il Remember that at present, the features that would be preselected by the semantics must be preselected by hand for each individual clause. However, this example illustrates the grammar we are working with, and demonstrates that this approach works very well with large grammars.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML