File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/88/c88-1006_metho.xml

Size: 15,826 bytes

Last Modified: 2025-10-06 14:12:08

<?xml version="1.0" standalone="yes"?>
<Paper uid="C88-1006">
  <Title>Morphology with Two-Level Rules and Negative Rule Features</Title>
  <Section position="4" start_page="0" end_page="28" type="metho">
    <SectionTitle>
2 Historical Note
</SectionTitle>
    <Paragraph position="0"> The formalism described here was developed with the goal of allowing the linguist to write rules with similar or even identical contexts and still have a way of processing them. This stands in contrast to Koskenniemi's formMism, which, in its initial formulation, seemed to rule out pairs of such rules.</Paragraph>
    <Paragraph position="1"> For instance, in Koskenniemi's formalism, as originally stated, the two rules below, a:bC/==a__~ a : c C/:==== o~__~, Would clash. Together they assert that a lexical character/a/, preceded by a sequence of character pairs c~ and followed by a sequence of character pairs fl, must correspond to both/b/and /c/on the surface.</Paragraph>
    <Paragraph position="2"> The orthographic rules described here are used in a morphological analysis system that is based on the work of Koskenniemi, Karttunen, and Wittenburg \[8,5\]. Its morphosyntactic component uses, instead of continuation classes, an extension of PATR  type rules including a device described by Karttunen \[4\] for handling disjunction. One version of this system also uses a definite-clause grammar in addition to the PATP~-type unification, and disjunction. It has been implemented in Prolog and runs on a Sun.</Paragraph>
  </Section>
  <Section position="5" start_page="28" end_page="29" type="metho">
    <SectionTitle>
3 Summary of Alternative Rule For-
</SectionTitle>
    <Paragraph position="0"> malism The basic idea behind the notion of two-level rule (due to Koskenniemi \[8\]) is that there are two levels of linguistic information to which a rule may refer. One has to do with how a morpheme is spelled in the lexicon. That is called the lexical level. The other has to do with how a morpheme appears in text, i.e., the surface representation. There is no way for rules to apply one after the other, creating and referring to intermediate levels of representation. Instead, rules are viewed as constraints On mappings between surface and underlying forms of morphemes. They stipulate how to get from underlying to surface form, and vice versa.</Paragraph>
    <Paragraph position="1"> Two-level rules in the alternative to Koskenniemi's formalism that I proposed in an earlier paper \[2\], take one of three forms:</Paragraph>
    <Paragraph position="3"> The a and ~ in the contexts of these rules represent strings of character pairs where one character of the pair refers to the lexical level of representation and the other refers to the surface.</Paragraph>
    <Paragraph position="4"> P~ule (1) is very similar to a standard phonological rule. It means roughly that lexical/a/ must correspond to surface/b/ in the context given. A more accurate and detailed description is as follows: if lexieal/a/occurs in tile given context, then it may not correspond to what it normally would correspond to, but it may correspond to surface/b/.</Paragraph>
    <Paragraph position="5"> tLule (2) means that lexical /a/ is allowed to correspond to surface/b/ in the context given, but not elsewhere. More precisely, the rule allows the pair /a:b/ (lexical /a/ corresponding to surface /b/) to occur in the context given and, unless there are other rules licensing the pair in other contexts, the context given is the only place where that correspondence is allowed.</Paragraph>
    <Paragraph position="6"> Rule (3) says that lexical/a/ may not correspond to surface /b/in the context given. Both rules (1)and (2) mention a character's default. A normal alphabetic character in this system defaults to itself. This means that a pair of alphabetic characters /a:a/ does not need to be licensed by a rule. In contrast to alphabetic characters (a through z)~ there are diacritic characters such as the plus sign (+) for morpheme boundaries. In Karttunen and Wittenburg's system, \[5\] there is also a backquote (') for representing stress; Koskenniemi uses several others as well, \[8\]. The default for lexical-level diacritics, at least in the system described here, is that they correspond to the null surface character, which is frequently written with a zero.</Paragraph>
    <Paragraph position="7"> .NC/,,atlve :~,uie Features There is a 9roblem with previous accounts of English that have been done in terms of two-level rules. There is no easy way to let the phonoh@cal rules know about individual idiosyncrasy in the lexical items. In the. work of Koskcnnicmi \[8\] and Karttunen and Wittenburg \[5\], diacritics are put into the lexical representation of a word in order to allow the linguist to write a phonological rule that applies in some words and not others according to the presence or absence of the diacritic. The diacritic is mentioned in the rule. The words that do not contain the diacritic do not undergo the rule.</Paragraph>
    <Paragraph position="8"> In oldofa~;hioned generative phonology, there was the notion of a negative rule feature to handle such cases. One could say of certMn :,norphemes ~hat appeared to be exceptions to certain phonological rules that such morphemes possessed a feature specifying that some partictflar phonological rule did not apply to themL The ide;~ of negative rule featmes has an adwmtage over the use of diacritics mentioned above in that it allows simplification of the photmlogicM rules and the lexicon. It seems to me more straightforward to have a lexical item that says miuus such and such a rule than to have the lexical item contain a colon or quotation marl&lt; whose function is to assert that some rule does not apply. The complexity of the lexical items is the same, but in the first case, at least, t'he phonological rule can be made simpler by omission of the arbitrary diacritic.</Paragraph>
    <Paragraph position="9"> There ar~, three examples from English orthography that will be used to help demonstrate how negative rule features may be employed.</Paragraph>
    <Paragraph position="10"> The anMysis of consonant gemination in Karttunen and Wittenburg's paper, \[5\], relies on the use of diacritics of just the sort mentioned above. A simplified versieiJ of the rule is given below. Gemin~ction: +:cl,Z~&gt;~ C* V -:el __V; where cl is in { b,d,f,g,l,m,n,p,r,s,t }.</Paragraph>
    <Paragraph position="11"> This rule uses a plus sign (+) for morpheme boundaries, and a backquote (') for accent where accent is important. It correctly describes the following data: questiouing versus *questionning, debiting versus *debitting, eating versus *catting.</Paragraph>
    <Paragraph position="12"> The rule also correctly describes the following data, provided the lexicM entr&gt; contains a backquote in the right place.</Paragraph>
    <Paragraph position="13"> referred versus. *refered (spellings in lexicon are &amp;quot;re'let&amp;quot; + &amp;quot;ed').</Paragraph>
    <Paragraph position="14"> In order to get the facts right for monosyllabic words, Karttunen and Wittenburg's rule also mentions that, instead of a backquote, a word boundary (#) will do.</Paragraph>
    <Paragraph position="15"> The only point of contention here is that their system requires the the lexicM entry to contain a diacritic (and furthermore the diacritic must be correctly located within the word). That the diacritic is reminiscent of an accent mark is no accident. Stress is clearly a fac,~or in English consonant gemination, q'helr solution is to find a way to represent stress in the orthography. The Mternative l:,roposed here is to express it in the form of a negative rule feature on the following sample iexical items. The rule is again simplified.</Paragraph>
    <Paragraph position="17"> where cl is in {b,d,f,g,l,m,n,p,r,s,t} Words: refer (default is that it is consistent with all rules) bother -gemination (means that the gemination rule does not apply to this word) There are other sets of data for which this technique is usefnl. The case that. comes to mind most readily deals with combining a noun o,&amp;quot; verb stem ending in/o/with an/s/morpheme representing, respectively, plural for nouns and third person singular for verbs. The following rules do well at describing these facts about English orthography.</Paragraph>
    <Paragraph position="18">  potato+s ==ee potatoes, *potatos (need an/e/) do+s --m~ does, *dos (need an/e/) piano+s &amp;quot;--::4, pianos, *l)ianocs (can't have an/e/) piccolo+s ==&gt; piccolos, *piccoloes (can't haw'. an/e/) banjo+s ==&gt; banjos or banjoes (both are acceptable) cargo+s =-~e- cargos or cargoes (both arc acceptable) The first of the epenthesis rules describes/potato+s/ ==&gt; \[potatoes\] and/do+s/==&gt; \[does\] correctly, but incorrectly states that the plural of/piano/ is */pianoes/. The second rule is weaker, generating all of the correct forms - but all of the wrong ones too, so that it achieves the right results for/banjo+s/ ==# \[banjoes\[ or \[banjos\] and likewise for/cargo+s/, but yields both the right and the wrong results for the others.</Paragraph>
    <Paragraph position="19"> The way to get the facts right is to put negative rule features on the lexicM items in question, as shown here:  The alternatives are either to list some forms as being irregular or to insert diacritics into some of the words so that the rule(s) will apply only to the correct lexical items. To list some of the forms as irregular is to miss the generalization that they are all irregular in exactly the same way. To use a diacritic (or possibly two) to describe the facts correctly may lead to making other, unrelated rules more complicated. Furthermore, it seems to be an attempt at expressing historical information, such as a word's provenance, in terms of abstract phonological segments.</Paragraph>
    <Paragraph position="20"> In general, the device of negative rule features seems to be well suited to tile task of passing information between a lexical entry and the phonology component. This is a useful capability. It is perhaps analogou,s to employing augmented phrase-structure rules in syntax when, at least in theory, pure context-free rules would do.</Paragraph>
    <Paragraph position="21"> The mMn idea here is that there is a way to let phonological (or orthographic) rules refer to features of a nmrpheme that may  not be easily represented as phonemic segements. As regards the gemination rule mentioned earlier, the right procedure might be to let the rule mention stress and store values for that feature in the lexical entries.</Paragraph>
  </Section>
  <Section position="6" start_page="29" end_page="29" type="metho">
    <SectionTitle>
5 Computer Interpretation of the
</SectionTitle>
    <Paragraph position="0"/>
    <Section position="1" start_page="29" end_page="29" type="sub_section">
      <SectionTitle>
Rules
</SectionTitle>
      <Paragraph position="0"> What makes these rules interesting is that there is a way to apply them in a morphological parser or generator. What follows is a description of the algorithm used by the code that I have implemented in Quintus Prolog on a Sun. When the rule epenthesisl is read in, it is decomposed into two rules. This rule, epenthesisl: +---&gt;e/O__s, yields these rules: epenthesis\]: +/e allowed in context o __ s epenthesis\]: +/0 disallowed in context o _ s.</Paragraph>
      <Paragraph position="1"> These rules are then stored as lists of character pairs: epenthesisl: allowed : o/o . +/e .s/.s epenthesisl: disallowed: o/o +/O s/s.</Paragraph>
    </Section>
  </Section>
  <Section position="7" start_page="29" end_page="29" type="metho">
    <SectionTitle>
7 Algorithm With Negative Rule
</SectionTitle>
    <Paragraph position="0"/>
    <Section position="1" start_page="29" end_page="29" type="sub_section">
      <SectionTitle>
Features
</SectionTitle>
      <Paragraph position="0"> Thus far, nothing has been said about how negative rule features enter into the picture. When a morpheme boundary is encountered, a morpheme has just been looked up in the lexicon. At that point, if it has some negative rule features on it, it is a simple matter to sort through the list of rules that have partially matched the input and discard those that the morpheme says do not apply. If that entails eliminating the last rule in some set of allowed-type rules that have all already matched past the main pair of the rule, then the input being scanned is not allowable as a possible mapping between lexical and surface forms. Otherwise one should just go on as before, comparing the rules with the input being scanned.</Paragraph>
    </Section>
  </Section>
  <Section position="8" start_page="29" end_page="29" type="metho">
    <SectionTitle>
8 Conclusion
</SectionTitle>
    <Paragraph position="0"> A general procedure for using phonological or orthographic two-level rules has been presented. These rules are much easier to refine and develop than automata transition tables. In addition, a method has been presented for listing which morphemes are exceptions to which \[orthographic\] rules, and an algorithm has been described that makes it possible to use this information in a straightforward way.</Paragraph>
    <Paragraph position="1"> Furthermore, these are two-level rules. As Koskenniemi has noted, \[8\], since these rules simply state correspondences between surface strings and underlying strings, they may be used either for doing generation or recognitio n. The device of negative rule features proposed here has the same power as Koskeniemi's device of putting arbitrary diacritics into selected classes of morpheines and rules, but is argued to be simpler.</Paragraph>
  </Section>
  <Section position="9" start_page="29" end_page="29" type="metho">
    <SectionTitle>
6 Basic Algorithm
</SectionTitle>
    <Paragraph position="0"> The rules are sequences of character pairs. A mapping between a string of lexical characters and a string of surface characters may also be considered to be a list of character pairs. No disallowedtype rule may be a substring of a mapping between a lexical string and a surface string.</Paragraph>
    <Paragraph position="1"> The rule checker proceeds down the list of character pairs, looking for any substring that is the same as one of the disallowedtype rules. If it finds one, the string of character pairs it was considering is not a valid mapping from a lexical form (word) to a surface form.</Paragraph>
    <Paragraph position="2"> The other type of rule, the allowed-type rule, is somewhat different. A dot is put into the rule right after the end of the left context to mark the next character pair as being the main pair of the rule. Any character pair that is the main pair for one of these allowed-type rules needs to be surrounded by the right and left contexts of one of these rules. The way that is checked for in this system is as follows. The string of charcter pairs is scanned from left to right. Each time a pair is encountered that is the same as the first pair of some allowed-type rules, the rules are put into a set. As more character pairs are scanned, they are compared with the sets of rules already encountered. Rules that do not continue to match the scanned input are ejected from the set.</Paragraph>
    <Paragraph position="3"> When the main pair of a rule in one of these sets is scanned, it is removed from the set it was in and put into a new one. The rules in this set are compared with scanned input in the same manner as before except that, if the last pair of some rule matches a pair that is being scanned, the whole set is discarded as no longer of interest. Conversely, if there is not at least one rule in the set that matches the scanned input all the way to the end, then the input being scanned is not an allowable mapping between lexical and surface forms.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML