File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/94/j94-3010_metho.xml

Size: 64,300 bytes

Last Modified: 2025-10-06 14:13:53

<?xml version="1.0" standalone="yes"?>
<Paper uid="J94-3010">
  <Title>Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems</Title>
  <Section position="2" start_page="457" end_page="459" type="metho">
    <SectionTitle>
3 The constraints proposed here deviate in various respects from the standard version of HPSG. We
</SectionTitle>
    <Paragraph position="0"> follow Carpenter (1992) in using the notation a =~ C/ to specify that type ~ satisfies constraint C/.</Paragraph>
    <Paragraph position="1"> 4 We are using what Carpenter (1992) calls TOTAL WELL-TYPING. That is, (i) the only attributes and values that can be specified for a given feature structure of type T are those appropriate for ~-; and (ii) every feature structure of type ~- must be specified for all attributes appropriate for T.</Paragraph>
    <Paragraph position="2">  Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems could be constrained to have the following features defined:</Paragraph>
    <Paragraph position="4"> That is, feature structures of type sign must contain the attributes PHON (i.e. phonology), SYNSEM (i.e. syntax/semantics), 5 and DTRS (i.e. daughters) and these attributes must take values of a specific type (i.e., phon, synsem, and list, respectively). A further crucial point is that appropriateness conditions are inherited by subtypes. For example, since morph is a subtype of sign, it inherits all the constraints obeyed by sign.</Paragraph>
    <Paragraph position="5"> Moreover, as we shall see in Section 3.2, it is subject to some further appropriateness conditions that are not imposed on any of its supertypes.</Paragraph>
    <Paragraph position="6"> Continuing in the same vein, we can assign appropriateness conditions to the types synsem and phon that occurred as values in (2), (simplifying substantially from standard HPSG). Here we give the constraints for synsem. The type phon will be discussed in Section 2.</Paragraph>
    <Paragraph position="8"> To conclude this section, we shall look very briefly at matters of interpretation and inference. As shown by Carpenter (1992) and Zajac (1992, in press), we can use constraint resolution to carry out type inference for feature terms. Following Zajac, let us say that a GROUND feature term is a term all of whose type symbols are minimal (i.e., the most specific types in the hierarchy immediately above _L). A WELL-TYPED feature term is one that obeys all the type definitions. Then the meaning of a feature term F is given by the set of all well-typed ground feature terms that are subsumed by F. Evaluating F, construed as a query, involves describing F's denotation; for example, enumerating all the well-typed ground feature terms it subsumes. Since the type deftnitions obeyed by F might be recursive, its denotation is potentially infinite. Consider for example the following definitions (where &amp;quot;nelist&amp;quot; and 'elist&amp;quot; stand for nonempty list and empty list respectively, and T subsumes every type): Example 4  a. list =~ nelist V elist b. FIRST : T \] nelist \[REST : list J 5 Earlier versions of HPSG kept syntax and semantics as separate attributes, and we shall sometimes revert to the latter when borrowing examples from other people's presentations.  Computational Linguistics Volume 20, Number 3 Here, the denotation of the type symbol list is the set of all possible ground lists. In practice, a constraint solver could recursively enumerate all these solutions; an alternative proposed by Zajac would be to treat the symbol LIST as the best finite approximation of the infinite set of all lists.</Paragraph>
    <Section position="1" start_page="459" end_page="459" type="sub_section">
      <SectionTitle>
1.4 Finite-State Phonology
</SectionTitle>
      <Paragraph position="0"> Over the last decade much has been written on the application of finite-state transducers (FSTs) to phonology, centering on the TWO-LEVEL MODEL of Koskenniemi (1983).</Paragraph>
      <Paragraph position="1"> Antworth (1990) and Sproat (1992) give comprehensive introductions to the field. The formalism is an attractive computational model for 1960s generative phonology. However, as has already been noted, phonologists have since moved away from complex string rewriting systems to a range of so-called nonlinear models of phonology. The central innovation of this more recent work is the idea that phonological representations are not strings but collections of strings, synchronized like an orchestral score. There have been some notable recent attempts to rescue the FST model from its linearity in order to encompass nonlinear phonology (Kay 1987; Kornai 1991; Wiebe 1992). However, from our perspective, these refinements to the FST model still admit unwarranted operations on phonological representations, as well as rule conspiracies and the like. Rather, we believe a more constrained and linguistically appealing approach is to employ finite-state automata (FSAs) in preference to FSTS, since it has been shown how FSAS can encode autosegmental representations and a variety of constraints on those representations (Bird and Ellison 1994). The leading idea in this work is that each tier is a partial description of a string, and tiers are put together using the intersection operation defined on FSAs.</Paragraph>
      <Paragraph position="2"> Apart from being truer to current phonological theorizing, this one-level model has a second important advantage over the two-level model. Since the set of FSAs forms a Boolean lattice under intersection, union, and complement (a direct consequence of the standard closure properties for regular languages), we can safely conjoin ('unify'), disjoin, and negate phonological descriptions. Such a framework is obviously compatible with constraint-based grammar formalisms, and there is no reason in principle to prevent us from augmenting HPSG with the data type of regular expressions. In practice, we are not aware of any existing implementations of HPSG (or other feature-based grammars) that accommodate regular expressions. Ideally, we would envisage a computational interpretation of typed feature structures where operations on regular expression values are delegated to a specialized engine that manipulates the corresponding FSAs and returns regular expression results. 6 This issue is discussed further in Section 6.</Paragraph>
    </Section>
    <Section position="2" start_page="459" end_page="459" type="sub_section">
      <SectionTitle>
1.5 Overview of the Paper
</SectionTitle>
      <Paragraph position="0"> The structure of the paper is as follows. In the next section, we present our assumptions about phonological representations and phenomena, couched in the framework of typed feature logic. In Section 3 we discuss our view of the lexicon, borrowing heavily on HPSG's lexical type hierarchy, and developing some operations and representations needed for morphology. The next two sections investigate various applications of the approach to two rather differing phenomena, namely Sierra Miwok templatic morphology and French schwa. Section 6 discusses some implementation issues. The paper concludes with a summary and a discussion of future prospects.</Paragraph>
    </Section>
  </Section>
  <Section position="3" start_page="459" end_page="460" type="metho">
    <SectionTitle>
6 A similar approach is envisaged by Krieger, Pirker, and Nerbonne (1993).
</SectionTitle>
    <Paragraph position="0"/>
  </Section>
  <Section position="4" start_page="460" end_page="467" type="metho">
    <SectionTitle>
2. String-Based Phonology
</SectionTitle>
    <Paragraph position="0"> In this section we present a string-based phonology based on the HPSG list notation.</Paragraph>
    <Paragraph position="1"> We present the approach in Section 2.1 and Section 2.2, concluding in Section 2.3 with a discussion of prosodic constituency.</Paragraph>
    <Section position="1" start_page="460" end_page="460" type="sub_section">
      <SectionTitle>
2.1 List Notations
</SectionTitle>
      <Paragraph position="0"> As a concession to existing practice in HPSG, we have taken the step of using lists in place of strings. We shall use angle bracket notation as syntactic sugar for the standard FIRST/REST encoding.</Paragraph>
      <Paragraph position="1"> We shall assume that the type system allows parameterized types of the form list(a), where a is an atomic type.</Paragraph>
      <Paragraph position="3"> We can now treat a* and a + as abbreviations for list(a) and nelist(a) respectively.</Paragraph>
      <Paragraph position="4"> Another useful abbreviatory notation is parenthesized elements within lists. We shall interpret (a(b)) --- L, a list consisting of an a followed by an optional b concatenated with an arbitrary list L, as the following constraint:</Paragraph>
      <Paragraph position="6"> We shall see applications of these list notations in the next section.</Paragraph>
    </Section>
    <Section position="2" start_page="460" end_page="460" type="sub_section">
      <SectionTitle>
2.2 A Prosodic Type Hierarchy
</SectionTitle>
      <Paragraph position="0"> A PROSODIC TYPE HIERARCHY is a subsumption network akin to the lexical hierarchy of HPSG (Pollard and Sag 1987). The type constraints we have met so far can be used to define a type hierarchy, which for present purposes will be a Boolean lattice. In this section we present in outline form a prosodic hierarchy that subsequent analyses will be based on. Example (7) defines the high-level types in the hierarchy.</Paragraph>
    </Section>
    <Section position="3" start_page="460" end_page="462" type="sub_section">
      <SectionTitle>
Example 7
</SectionTitle>
      <Paragraph position="0"> phon ~ utterance v phrase V foot v syl V segment Each of these types may have further structure. For example, following Clements (1985:248) we may wish to classify segments in terms of their place and manner of articulation, using the following appropriateness declaration.</Paragraph>
      <Paragraph position="2"> Suppose now that we wished to use these structures in a constraint for English homorganic nasal assimilation. This phenomenon does not occur across phonological phrase boundaries and so the constraint will be part of the definition of the type (phonological) phrase. Let us assume that a phrase is equivalent to segment*, i.e. a list of segments. Informally speaking, we would like to impose a negative filter that bars any nasal whose value for place of articulation differs from that of the stop consonant that immediately follows. Here, we use SL as an abbreviation for SUPRALARYNGEAL, CONT for CONTINUANT, MN for MANNER, and PL for PLACE.</Paragraph>
      <Paragraph position="3"> Example 9 ~phrase I'&amp;quot;segment \[SL : \[ c d PL : segment \[SL : PL : ' While the abbreviatory conventions in this filter might appear suspicious, it is straight-forwardly translated into the constraint in (10). This constraint is divided into three parts. The first simply requires that hna be a subtype of list(segment). The second part is lifted from (9), ensuring that the first two positions in the list do not violate the assimilation constraint. The third part propagates the assimilation constraint to the rest of the list.</Paragraph>
      <Paragraph position="5"> Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems Standard techniques can now be used to move the negation in (10) inward. 7 Since constraints on adjacent list elements generally seem to be more intelligible in the format exhibited by (9), we shall stick to that notation in the remainder of the paper.</Paragraph>
    </Section>
    <Section position="4" start_page="462" end_page="463" type="sub_section">
      <SectionTitle>
2.3 Prosodic Constituency
</SectionTitle>
      <Paragraph position="0"> One standard phonological approach assumes that prosodic constituency is like phrase structure (Selkirk 1984). For example, one might use a rewrite rule to define a (phonological) phrase as a sequence of feet, and a foot as sequence of syllables: Example 11 a. phrase ~ foot + b. foot --* syl + Within the framework of HPSG, it would be simple to mimic such constituency by admitting a feature structure of type phrase whose DTRS (i.e. daughters) are a list of feature structures of type foot, and so on down the hierarchy. However, there appears to be no linguistic motivation for building such structure. Rather, we would like to say that a phrase is just a nonempty list of feet. But a foot is just a list of syllables, and if we abandon hierarchical structure (e.g. by viewing lists as strings), we seem to be stuck with the conclusion that phrases are also just lists of syllables. In a sense this is indeed the conclusion that we want. However, not any list of syllables will constitute a phrase, and not every phrase will be a foot. That is, although the data structure may be the same in each case, there will be additional constraints that have to be satisfied. For example, we might insist that elements at the periphery of phrases are exempt from certain sandhi phenomena; and similarly, that feet have no more than three syllables, and only certain combinations of heavy and light syllables are permissible. Thus, we shall arrive at a scheme like the following, where the Ci indicate the extra constraints: 8 Example 12 a. phrase =_ foot + A C1 A ... A Ck b. foot = syl + ACt A ... A Cn This concludes our discussion of string-based phonology. We have tried to show how a phonological model based on FSAs is compatible with the list notation and type regime of HPSG. Next we move onto a consideration of morphology and the lexicon. 7 These techniques employ the following equivalences: ~\[A: ~\] = -~\[A: ~\]V~\[B: @\] -~\[A: q~\] = \[-~(A: T)\]V\[A: ~*\] Here ~(A:T) indicates that the attribute A is not appropriate for this feature structure. 8 Sproat and Brunson (1987) have also proposed a model in which prosodic constituents are defined as conjunctions of constraints.</Paragraph>
      <Paragraph position="1">  Computational Linguistics Volume 20, Number 3 3. Morphology and the Lexicon</Paragraph>
    </Section>
    <Section position="5" start_page="463" end_page="464" type="sub_section">
      <SectionTitle>
3.1 Linguistic Hierarchy
</SectionTitle>
      <Paragraph position="0"> The subsumption ordering over types can be used to induce a hierarchy of grammatically well-formed feature structures. This possibility has been exploited in the HPSG analysis of the lexicon: lexical entries consist of the idiosyncratic information particular to the entry, together with an indication of the minimal lexical types from which it inherits. To take an example from Pollard and Sag (1987), the base form of the English verb like is given in Example 13.</Paragraph>
      <Paragraph position="2"> Since main is a subtype of verb, the entry for like will inherit the constraint that its major class feature is V; by virtue of the type strict-trans, it will inherit the constraint that the first element in the SUBCAT list is an accusative NP, while the second element is a nominative NP; and so on for various other constraints. Figure 1 shows a small and simplified portion of the lexical hierarchy in which the verb like is a leaf node.</Paragraph>
      <Paragraph position="3"> Along the phonological dimension of signs, lexical entries will have to observe any morpheme or word level constraints that apply to the language in question. When words combine as syntactic phrases, they will also have to satisfy all constraints on well-formed phonological phrases (which is not to say that phonological phrases are isomorphic with syntactic ones). In the general case, we may well want to treat words in the lexicon as unsyllabified sequences of segments. It would then follow that, for example, the requirement that syllable-initial voiceless obstruents be aspirated in English  A portion of the lexical hierarchy.</Paragraph>
      <Paragraph position="4">  Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems would have to be observed by each syllable in a phrase (which in the limiting case, might be a single word), rather than lexical entries per se.</Paragraph>
      <Paragraph position="5"> In some languages we may require there to be a special kind of interaction between the lexical and the prosodic hierarchy. For example, Archangeli and Pulleyblank (1989) discuss the tongue root harmony of Yoruba, which is restricted to nouns. If atr (i.e. advanced tongue root) was the type of harmonic utterances, then we could express the necessary constraint thus:</Paragraph>
      <Paragraph position="7"> This kind of constraint is known as a morpheme structure constraint, and phonologists have frequently needed to have recourse to these (Kenstowicz and Kisseberth 1979). Another interaction between prosody and morphology is the phenomenon of prosodic morphology, an example of which can be found in Section 4.</Paragraph>
    </Section>
    <Section position="6" start_page="464" end_page="466" type="sub_section">
      <SectionTitle>
3.2 Morphological Complexity
</SectionTitle>
      <Paragraph position="0"> Given the syntactic framework of HPSG, it seems tempting to handle morphological complexity in an analogous manner to syntactic complexity. That is, morphological heads would be analyzed as functors that subcategorize for arguments of the appropriate type, and morphemes would combine in a Word-Grammar scheme. Simplifying drastically, such an approach would analyze the English third person singular present suffix -s in the manner shown in (15), assuming that affixes are taken to be heads.</Paragraph>
      <Paragraph position="2"> By adding appropriately modified versions of the Head Feature Principle, Subcategor- null ization Principle, and linear order statements, such a functor would combine with a verb stem to yield a tree-structured sign for walks.</Paragraph>
      <Paragraph position="3"> Example 16 \[P ON: EP ON {ver _s e \[PHON verb While one may wish to treat derivational morphology in this way (cf. Krieger  Note that we have added to sign a new attribute MORPH, with a value morph. The latter has two subtypes, affix-morph and basic-morph, depending on whether the value contains a stem and affix or just a stem.</Paragraph>
      <Paragraph position="4"> Example 18 morph ~ affix-morph V basic-morph While both of these types will inherit the attribute STEM, affix-morph must also be defined for the attribute AFFIX:  Example 19 a. \[STEM:stem\] morph b. affix_morph \[ AFFIx:affix\] Moreover, affix has two subtypes: Example 20 affix ~ prefix V suffix Thus, (17) is a third person singular verb form whose stem is unspecified.  As indicated in Section 1.3, we can take the interpretation of a complex type to be equivalent to the disjunction of all of its subtypes. Now, suppose that our lexicon contained only two instances of verb-stems, namely walk and meet. Then (17) would evaluate to exactly two fully specified word forms, where verb-stem was expanded to the signs for walk and meet respectively. Example 21 illustrates the first of these options.</Paragraph>
      <Paragraph position="6"> &amp;quot;PHON : \[~(w OI k} \] SYNSEM: \[CAT: verb\]\] verb-stem suffix \[PHON : \[\]\[\] (s} 1  Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems Of course, this statement of suffixation would have to be slightly enriched to allow for the familiar allomorphic alternation -s,,~z,,,~z. The first pair of allomorphs can be handled by treating the suffix as unspecified for voicing, together with a voicing assimilation rule similar to the homorganic nasal rule in (9). The third allomorph would admit an analysis similar to the one we propose for French schwa in Section 5. A second comment on (21) is that the information about ordering of affixes relative to the stem should be abstracted into a more general pair of statements (one for prefixes and one for suffixes) that would apply to all morphologically complex lexical signs (e.g. of type affixed); this is straightforward to implement:  Given this constraint, it is now unnecessary to specify the phonology attribute for feature terms like (21).</Paragraph>
      <Paragraph position="7"> Additionally, it is straightforward to prevent multiple copies of the plural suffix from being attached to a word by ensuring that 3ps and verb-stem are disjoint.</Paragraph>
    </Section>
    <Section position="7" start_page="466" end_page="467" type="sub_section">
      <SectionTitle>
3.3 Morphophonological Operations
</SectionTitle>
      <Paragraph position="0"> In and of itself, HPSG imposes no restrictions on the kind of operations that can be performed in the course of composing morphemes into words, or words into phrases.</Paragraph>
      <Paragraph position="1"> As an illustration, consider the data from German verb inflections analyzed by Krieger, Pirker, and Nerbonne (1993). As they point out, the second person singular present inflection -st has three different allomorphs, phonologically conditioned by the stem: Example 23 sag+st arba~t+0st m~ks+t 'say' 'work' 'mix' Although the main thrust of their paper is to show how an FST treatment of this allomorphy can be incorporated into an HPSG-style morphological analysis, from a purely formal point of view, the FST is redundant. Since the lexical sign incorporates the phonologies of both stem and affix, segments can be freely inserted or deleted in constructing the output phonology. This is exemplified in (24) for arbeitest and mixt respectively.</Paragraph>
      <Paragraph position="3"> That is, we can easily stipulate that o is intercalated in the concatenation of stem and suffix if the stem ends with a dental stop (i.e either t or d); and that the s of the suffix is omitted if the stem ends with alveolar or velar fricative. Although an actual analysis along these lines would presumably be stated as a conditional, depending on the form of the stem, the point remains that all the information needed for manipulating the realization of the suffix (including the fact that there is a morpheme boundary) is already available without resorting to two level rules} deg Of course, the question this raises is whether such operations should be permitted, given that they appear to violate the spirit of a constraint-based approach. The position we shall adopt in this paper is that derivations like (24) should in fact be eschewed. That is, we shall adopt the following restriction:</Paragraph>
    </Section>
    <Section position="8" start_page="467" end_page="467" type="sub_section">
      <SectionTitle>
Phonological Compositionality:
</SectionTitle>
      <Paragraph position="0"> The phonology of a complex form can only be produced by either unifying or concatenating the phonologies of its parts.</Paragraph>
      <Paragraph position="1"> We believe that some general notion of phonological compositionality is methodologically desirable, and we assume that Krieger, Pirker, and Nerbonne would adopt a similar position to ours. The specific formulation of the principle given above is intended to ensure that information-combining operations at the phonological level are monotonic, in the sense that all the information in the operands is preserved in the result. As we have just seen, the constraint-based approach does not guarantee this without such an additional restriction.</Paragraph>
    </Section>
  </Section>
  <Section position="5" start_page="467" end_page="474" type="metho">
    <SectionTitle>
4. Sierra Miwok Templatic Morphology
</SectionTitle>
    <Paragraph position="0"> Noncatenative morphology has featured centrally in the empirical motivations for autosegmental phonology, since McCarthy's demonstration that the intercalation of vowels in Arabic consonantal verb roots could be elegantly handled within this framework (McCarthy 1981). This section presents an approach to intercalation that uses key 10 This approach of using restructuring devices in the process of a derivation has been explored in the context of extended Montague frameworks by Wheeler (1981) and Hoeksema and Janda (1988).</Paragraph>
    <Paragraph position="1">  Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems insights from autosegmental phonology. However, they are captured within constraint-based grammar where the inflectional paradigm is realized as an inheritance hierarchy of partially instantiated stem forms (cf. Reinhard and Gibbon \[1991\]). We also show that autosegmental association of consonants and vowels to a skeleton can be modeled by reentrancy. Rather than classical Arabic, we use the simpler data from Sierra Miwok that Goldsmith (1990) chose to illustrate the phenomenon of intercalation in his textbook.</Paragraph>
    <Paragraph position="2"> This section is divided into four subsections. In Section 4.1 we present an overview of the data, and in Section 4.2 we briefly show what a traditional generative analysis might look like. Our encoding of association by reentrancy is given in Section 4.3, while Section 4.4 contains our constraint-based analysis of Sierra Miwok stem forms.</Paragraph>
    <Section position="1" start_page="468" end_page="469" type="sub_section">
      <SectionTitle>
4.1 Descriptive Overview
</SectionTitle>
      <Paragraph position="0"> As mentioned above, Goldsmith (1990) takes data from Sierra Miwok verb stems to illustrate morphologically determined alternations in skeletal structure. He discusses three of the four types of stem, where the division into types depends primarily on the syllable structure of the basic form, which is the form used for the present tense. The three types are given the following autosegmental representations by Gold- null As shown in (26), each type has forms other than the basic one, depending on the morphological or grammatical context; these additional forms are called second, third, and fourth stems.</Paragraph>
      <Paragraph position="1"> Although the associations of vowels and consonants exhibited above are taken as definitional for the three stem Types, from the data in (26) it appears that the distinction is only relevant to so-called Basic stem forms.</Paragraph>
      <Paragraph position="2">  bleed kicaaw kicaww kiccaw kicwa jump tuyaa~ tuyar3r 3 tuyya U tuyr3a take patiit patitt pattit patti roll huteel hutell huttel hutle Type II quit celku celukk celluk celku go home wo?lu wo?ull wo??ul wo?lu catch up with nakpa nakapp nakkap nakpa spear wimki wimikk wimmik wimki Type III bury hamme hame?? hamme? ham?e dive ?uppi ?upi?? ?uppi~ ~up?i speak liwwa liwa?? liwwa? liw?a sing milli mili?? milli? mil?i</Paragraph>
    </Section>
    <Section position="2" start_page="469" end_page="470" type="sub_section">
      <SectionTitle>
4.2 Segmental Analysis
</SectionTitle>
      <Paragraph position="0"> Goldsmith (1990) has shown just how complex a traditional segmental account of Sierra Miwok would have to be, given the assumption that all of the stem forms are derived by rule from a single underlying string of segments (e.g. that kicaww, kiccaw and kicwa are all derived from kicaaw). Here, we simplify Goldsmith's analysis so that it just works for Type I stems. The left-hand column of (27) contains four rules, and these are restricted to the different forms according to the second column.</Paragraph>
      <Paragraph position="1">  Thus, the first rule requires that a vowel Vi is deleted if it occurs after a consonant and immediately before an identical vowel Vi that in turn is followed by a stem-final consonant. Goldsmith soundly rejects this style of analysis in favor of an autosegmental one: This analysis, with all its morphologically governed phonological rules, arbitrary rule ordering, and, frankly, its mind-boggling inelegance, ironically misses the most basic point of the formation of the past tense in Sierra Miwok. As we have informally noted, all the second  Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems stern forms are of the shape CVCVCC, with the last consonant a geminate, and the rules that we have hypothetically posited so far all endeavor to achieve that end without ever directly acknowledging it.</Paragraph>
      <Paragraph position="2"> (Goldsmith 1990:87)</Paragraph>
    </Section>
    <Section position="3" start_page="470" end_page="470" type="sub_section">
      <SectionTitle>
4.3 Association
</SectionTitle>
      <Paragraph position="0"> We shall not attempt here to give a general encoding of association, although the technique used in Sections 5.4 could be applied to achieve this end. Moreover, like Goldsmith we shall ignore the role of syllable structure in the data, though it clearly does play a role. Instead, we shall confine our attention to the manner in which skeletal slots are linked to the consonant and vowel melodies. Consider again the skeletal structure of Type I verb stems shown in (25a). As Goldsmith (1990) points out, there is a closely related representation that differs only in that the CV information is split across two tiers (and which allows a much more elegant account of metathesis and</Paragraph>
    </Section>
    <Section position="4" start_page="470" end_page="474" type="sub_section">
      <SectionTitle>
4.4 Basic Stem Forms
</SectionTitle>
      <Paragraph position="0"> The analysis starts from the assumption that the Sierra Miwok lexicon will contain minimally redundant entries for the three types of verb root. Let us consider the root corresponding to the basic stem form kicaaw. We take the unpredictable information to be the consonantal and vowel melodies, the valency, the semantics, and the fact it is a Type I verb stem. This is stated as (30), together with the declaration that lex-bleed is a subtype of v-root-I.</Paragraph>
      <Paragraph position="1">  Notice that we have said nothing about how the melodies are anchored to a skeleton--this will be a task for the morphology. Additionally, this entry will inherit various properties by virtue of its type v-root-I. The three types of verb root share at least one important property, namely that they are all verbs. This is expressed in the next two declarations:</Paragraph>
      <Paragraph position="3"> We shall also assume, for generality, that every v-root is a root, and that every root is a morph. Anticipating the rest of this section, we show how all the postulated types are related in Figure 2. The next step is to show how a v-root-I like (30) undergoes morphological modification to become a basic verb stem; that is, a form with skeletal structure. Our encoding of the morphology will follow the lines briefly sketched in Section 3.2.</Paragraph>
      <Paragraph position="4"> We begin by stating types that encode the patterns of skeletal anchoring associated with the three types of basic stem.</Paragraph>
      <Paragraph position="5">  Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems The appropriateness constraints on these types are given in (33). As an aid to readability, the numerical tags are supplemented with a C or a V to indicate the type of value involved.</Paragraph>
      <Paragraph position="7"> Each of these types specializes the constraints on the type phon, and each can be unified with the phon value earlier assigned to the root form of kicaaw in (30). In particular, the conjunction of constraints given in (34) evaluates to (29), repeated here:  However, we also need to specify the dependency between the three types of verb root, and the corresponding phonological exponents that determine the appropriate basic stem forms (cf. Anderson \[1992\]). As a first attempt to express this, let us say that stem can be either basic or affixed:  Computational Linguistics Volume 20, Number 3 We could now disjunctively specify the following three sets of constraints on basic:  Although the example in question does not dramatize the fact, this manner of encoding morphological dependency is potentially very redundant, since all the common constraints on basic have to be repeated each time. n In this particular case, however, it is easy to locate the dependency in the phon value of the three subtypes of v-root, as follows:</Paragraph>
      <Paragraph position="9"> We then impose the following constraint on basic:</Paragraph>
      <Paragraph position="11"> By iterating through each of the subtypes of v-root, we can infer the appropriate value of PHON within MORPH'S ROOT, and hence infer the value of PHON at the top level of the feature term. Example 40 illustrates the result of specializing the type v-root to lex-bleed: 11 In an attempt to find a general solution to this problem in the context of German verb morphology, Krieger and Nerbonne (in press) adopt the device of 'distributed disjunction' to iteratively associate morphosyntactic features in one list with their corresponding phonological exponents in another list.</Paragraph>
      <Paragraph position="12">  Exactly the same mechanisms will produce the basic stem for the other two types of verb root. For an account of the other alternations presented in Goldsmith's paradigm, and for some discussion of how lexical and surface forms determine each other, see Klein (1993).</Paragraph>
      <Paragraph position="13"> We have just seen an application of constraint-based phonology to Sierra Miwok.</Paragraph>
      <Paragraph position="14"> In order to illustrate some of the other expressive capabilities of the approach, we now turn to the phenomenon of French schwa.</Paragraph>
    </Section>
  </Section>
  <Section position="6" start_page="474" end_page="485" type="metho">
    <SectionTitle>
5. French Schwa
</SectionTitle>
    <Paragraph position="0"> Many phonological alternations can be shown to depend on properties of prosodic structure. In this section we show how the French phenomenon of schwa-zero alternation arises out of the interplay of various syllable structure requirements. This is done by introducing a system of prosodic types for syllables and a special type declaration showing how a string of segments can be 'parsed' into syllables. The standard (but nonmonotonic) ONSET MAXIMIZATION PRINCIPLE is reinterpreted in the system, as well as the exceptions to this principle due to a class of words known as h-aspir6 words. We also show how a certain kind of disjunction can be used to deal with free variation. As we shall see, some linguistic analyses are more amenable to a declarative encoding than others. In order to demonstrate this, it will be necessary to go into some detail concerning the linguistic data.</Paragraph>
    <Paragraph position="1"> This section is divided into four subsections. In Section 5.1 we present a descriptive overview of the data, 12 and in Section 5.2 we sketch a traditional generative analysis. A more recent, nonlinear analysis appears in Section 5.3 while our own, constraint-based version is presented in Section 5.4.</Paragraph>
    <Section position="1" start_page="474" end_page="476" type="sub_section">
      <SectionTitle>
5.1 Descriptive Overview
</SectionTitle>
      <Paragraph position="0"> Unlike schwa in English, the French schwa (or mute e) is a full vowel, usually realized as the low-mid front rounded vowel ce (and sometimes as the high-mid front rounded vowel o in certain predictable environments). Its distinctive characteristic is that under 12 The data is from standard French taken from (cited) literature, although in some instances we have found speakers with different acceptability judgments than reported here. See Morin (1987) for a discussion of some problems with the treatment of French data in the literature.</Paragraph>
      <Paragraph position="1">  Computational Linguistics Volume 20, Number 3 certain conditions, it fails to be realized phonetically. 13 From now on we shall use the term 'schwa' to refer to the vowel with this characteristic, rather than to the segment o. Although schwa is associated witlh orthographic e, not all es will concern us here.</Paragraph>
      <Paragraph position="2"> For example, the orthographic e of samedi \[sam.di\] 'Saturday' can be taken to indicate that the previous vowel should not be nasalized, while the final e of petite \[poe.tit\] indicates that the final t should be pronounced. In morphology, orthographic e marks feminine gender, first-conjugation verbs, and subjunctive mood.</Paragraph>
      <Paragraph position="3"> Instead, we shall be concerned with the pattern of realization and non-realization exhibited by schwa--a pattern that we interpret as grounded in the alternation of two allophones of schwa: oe and 0 (zero). This alternation is manifested in forms like (41), 14 where the dots indicate syllable boundaries.</Paragraph>
      <Paragraph position="4"> Example 41 a. six melons \[si.moe.15\] ,,~ \[sim.15\] b. sept melons \[s~t.moe.15\], *\[s~tmlS\] Observe that while six melons can be pronounced with or without the schwa, sept melons requires the schwa in order to break up the tml cluster that would otherwise be formed. Unfortunately, the conditions on the distribution of schwa are not as simple (and purely phonological) as this example implies. As we shall see, schwa alternation in French is governed by an interesting mixture of lexical and prosodic constraints.</Paragraph>
      <Paragraph position="5"> In the remainder of this section, we dispel the initial hypothesis that arises from (41), namely that schwa alternation is to be treated as a general epenthesis process, is Consider the following data (Morin 1978:111).</Paragraph>
      <Paragraph position="6">  The table in (42) gives data for the clusters \[rdr\], \[r f\], \[skl\] and \[ps\]. In the first column of data, the oe is possible or obligatory, while in the second column, it is absent. Thus, we see that the apperance of oe cannot be predicted on phonotactic grounds alone. Consequently, we shall assume that schwa must be encoded in lexical representations. Note that it is certainly not the case that a lexical schwa will be posited wherever there is an orthographic e. Consider the data in (43), where these orthographic es are underlined.</Paragraph>
      <Paragraph position="7"> 13 The data used in this section is drawn primarily from the careful descriptive work of Morin (1978) and Tranel (1987b). The particular approach to French schwa described in the following paragraphs most closely resembles the analysis of Tranel (1987a).</Paragraph>
      <Paragraph position="8">  In a purely synchronic analysis there is no basis for discussing an alternating vowel for bordereau, fais-lC/ and pelleterie. Many orthographic es that are not in the first syllable of a word come into this category.</Paragraph>
      <Paragraph position="9"> Accordingly, we begin our analysis with three background assumptions: the alternating schwa is (i) prosodically conditioned, (ii) lexically conditioned, and (iii) not in direct correspondence with orthographic e. Next we present a generative analysis of schwa due to Dell, followed by an autosegmental analysis due to Tranel. We conclude with our own, syllable-based analysis.</Paragraph>
    </Section>
    <Section position="2" start_page="476" end_page="477" type="sub_section">
      <SectionTitle>
5.2 A Traditional Generative Analysis
</SectionTitle>
      <Paragraph position="0"> The traditional approach to vowel-zero alternations is to employ either a rule of epenthesis or a deletion rule. Dell discusses the case of the word secoue, whose pronunciation is either \[sku\] or \[soeku\], in a way that parallels (41).</Paragraph>
      <Paragraph position="1"> In order to account for alternations such as that between \[sku\] and \[soeku\] there are two possibilities: the first consists of positing the underlying representation /sku/ where no vowel appears between /s/ and /k/, and to assume that there exists a phonological rule of epenthesis that inserts a vowel ce between two consonants at the beginning of a word when the preceding word ends in a consonant ....</Paragraph>
      <Paragraph position="2"> The second possibility is preferable: the vowel \[oe\] that appears in Jacques secoue is the realisation of an underlying vowel/o/which can be deleted in certain cases. We shall posit the VCE1 rule, which deletes any /a/ preceded by a single word-initial consonant when the preceding word ends in a vowel.</Paragraph>
      <Paragraph position="3"> VCEI: 0--+ 0 / V &amp;quot;#1 C (Dell, 1980:161f) Suppose we were to begin our analysis by asking the question: how are we to express the generalization about schwa expressed in the above rule? Since our declarative, monostratal framework does not admit deletion rules, we would have to give up. As we shall see below, however, we begin with a different question: how are we to express the observation about the distribution of schwa that Dell encodes in the above rule? There is another good reason for taking this line. As it happens, there is an empirical problem with the above rule, which Dell addresses by admitting a potentially large number of lexical exceptions to the rule and by making ad hoc stipulations (Dell 1980). Additionally, adding diacritics to lexical entries to indicate which rules they undergo and employing rules that count # boundaries would seem to complicate a grammar formalism unnecessarily. As we saw above for the discussion of the word bordereau, in the approach taken here we have the choice between positing a stable  Computational Linguistics Volume 20, Number 3 oe or one that alternates with zero (i.e. a schwa) in the lexicon, whereas Dell must mark lexical items to indicate which rules they must not undergo. There is also some evidence for a distinction between the phonetic identity of the ce allophone of schwa and the phonetic identity of a nonalternating lexical oe in some varieties of French, requiring that the two be distinguished phonologically (Morin 1978).</Paragraph>
      <Paragraph position="4"> Thus, the fact that Dell's analysis involves deletion does not provide a significant stumbling block to our approach. However, Dell employs another procedural device, namely rule ordering, in the application of the rule. In discussing the phrase vous me le dites \[vu.m(oe).l(oe).dit\], in which either schwa (but not both) may be omitted, Dell writes: VCE1 begins on the left and first deletes the schwa of me, producing /vu#m#1o#dit/. But VCE1 cannot operate again and delete the schwa of le, for, although this schwa was subject to the rule in the original representation, it no longer is once the schwa of me has been dropped.</Paragraph>
      <Paragraph position="5"> In other words, the first application of VCE1 creates new conditions that prevent it from operating again in the following syllable (Dell 1980:228).</Paragraph>
      <Paragraph position="6"> Again, we are not interested in encoding Dell's particular generalization, and in fact we are unable to. Rather, it is necessary to look at the underlying observation about the distribution of schwa. The observation is that schwa does not appear as its zero allophone in consecutive syllables. This observation is problematic for us, in that it refers to two levels of representation, an underlying (or lexical) level involving a schwa segment, and a surface level involving a zero allophone. We cannot formulate this observation monostratally. However, we can come up with a different observation, namely that the vowel is never omitted if the omission results in unacceptable syllable structure. In the case of Dell's example, vous me le dites, if both schwas are omitted the result is a \[vml\] cluster, which cannot be broken up into a valid coda-onset sequence. This new observation makes a different empirical prediction, namely that schwa can be omitted in consecutive syllables just in case the result is syllabifiable. As we shall see below in (51), this prediction is actually borne out.</Paragraph>
      <Paragraph position="7"> Before proceeding with our own analysis, we present an overview of an autosegmental analysis of French schwa due to Tranel. This analysis is interesting because it demonstrates the oft-repeated phenomenon of enriched representations leading to dramatically simplified rule systems. Given the heavy restriction on rules in a mono-stratal framework, it will be more natural to take Tranel's (rather than Dell's) analysis as our starting point.</Paragraph>
    </Section>
    <Section position="3" start_page="477" end_page="479" type="sub_section">
      <SectionTitle>
5.3 Tranel's Analysis
</SectionTitle>
      <Paragraph position="0"> Tranel (1987a) provides an insightful analysis of French schwa cast in the framework of autosegmental phonology. In this section we give an overview of this analysis. In the following section we shall endeavour to provide an empirically equivalent analysis.</Paragraph>
      <Paragraph position="1"> Tranel adopts a CV skeleton tier and a segmental tier. Schwa is represented as an unlinked vowel, as shown in the following representation for melons.</Paragraph>
      <Paragraph position="3"> Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems On top of this two-tiered structure, Tranel proposes a level of hierarchical organization for representing syllable structure. Tranel adopts the two syllable formation rules given in (45). A third (unstated) rule is assumed to incorporate unsyllabified consonants into coda position.</Paragraph>
      <Paragraph position="4"> Example 45 a. Basic syllable formation b. Schwa-syllable formation</Paragraph>
      <Paragraph position="6"> Note that (45a) does not apply to the mo~ sequence in (44), as the schwa is not linked to a V node as required on the left-hand side of rule (45a). (Tranel later adopts a refinement to (45a), preventing it from applying if the V is the first vowel of an h-aspir6 morpheme.) For the phrases six melons and sept melons, the basic syllable formation rule builds the following structures.</Paragraph>
      <Paragraph position="8"> The remaining consonants must either be syllabified leftward into an unsaturated coda or remain unsyllabified and rescued by the schwa syllable formation rule. For six melons, both options are possible, as illustrated below. Note that the unlinked oe is assumed to be phonetically uninterpreted.</Paragraph>
      <Paragraph position="10"> ce 1 5 s i m oe I 5 This gives us the two options, \[sim.lS\] and \[si.mce.15\], according with the observation  Computational Linguistics Volume 20, Number 3 in (41). For sept melons, however, there is just the one option. The t must be syllabified into the preceding coda, and the m requires the presence of schwa, and so we have \[set.mce.15\]. Further examples of this particular kind of schwa alternation are given</Paragraph>
    </Section>
    <Section position="4" start_page="479" end_page="480" type="sub_section">
      <SectionTitle>
Schwa Optional
</SectionTitle>
      <Paragraph position="0"> vous parlez de qui? ne te casse pas la t~te il est debout c'est depuis quatre ans  So far, we have seen the case where the leftward syllabification of a consonant licenses the omission of schwa. Now we turn to a similar case, but where the consonant syllabifies rightward into a following onset provided that the resulting onset cluster is permitted. The data in (49) are from Tranel (1987b:92).  Rule (50a) incorporates as many consonants as possible into an onset so long as the onset conforms to the phonotactic constraints of the language. Rule (50b), of most interest here, allows for a consonant to be incorporated into a following onset even if there is an intervening schwa, provided that the consonant is word-initial (and that the resulting onset is allowable). The intervening schwa remains unpronounced. Rule (50b), which is optional, correctly captures the alternations displayed in (49). This rule is restricted to apply word-initially &amp;quot;so as to avoid the generation of wordqnternal triliteral consonant clusters from underlying/CCoC/sequences (compare marguerite  Steven Bird and Ewan Klein Phonological Analysis in Typed Feature Systems /margorit/ \[margorit\] *\[margrit\] and margrave /margrav/ \[margrav\] *\[margorav\])&amp;quot; (Tranel 1987a:852). Thus, although many CCC sequences are acceptable phonologically, they are not permitted if a schwa is available to break up the cluster. We also note that Tranel's analysis (Tranel 1987a) gives the correct result for cases of deletion of schwa in consecutive syllables. Consider the following data. Example 51 a. on ne se moque pas \[6n.sm3k.pct\] (Valdman 1976:120) b. sur le chemin \[syl.fm~\] (Morin 1978:82) For both of these cases we observe an &amp;quot;underlying&amp;quot; ClceC2ce pattern, but where both ces are omitted and where C1 syllabifies into the preceding coda and C2 syllabifies into the following onset.</Paragraph>
      <Paragraph position="1"> To conclude, we can summarize the empirical content of Tranel's analysis as follows:  (a) Every consonant must be syllabified.</Paragraph>
      <Paragraph position="2"> (b) Schwa must be realized if it provides the syllable nucleus for an immediately preceding consonant that: (i) cannot be syllabified into a coda, and (ii) cannot form a permissible (word) onset with an immediately  following consonant.</Paragraph>
      <Paragraph position="3"> Naturally, this statement is not the last word on French schwa and there may be ways in which it needs to be revised, such as for the treatment of word-final schwas and thematic schwas (Tranel 1987a:855ff). However, since our purpose is primarily to illustrate the workings of the theoretical model, we shall take the above statement as a well-defined starting point on which to base the following analysis.</Paragraph>
    </Section>
    <Section position="5" start_page="480" end_page="485" type="sub_section">
      <SectionTitle>
5.4 A Constraint-Based Analysis
</SectionTitle>
      <Paragraph position="0"> Given our formal semantics for the autosegmental notation, it would be a relatively straightforward matter to implement Tranel's analysis directly, especially since the rules only involve the building of structure, and there is no use of destructive processes.</Paragraph>
      <Paragraph position="1"> Tranel's analysis is fully declarative.</Paragraph>
      <Paragraph position="2"> However, as it happens, there is no need for us to adopt the rich representation Tranel employs. We can simulate his analysis using a single tier (rather than two) while retaining a representation of syllable structure. Observe that the use of the CV tier and the melody tier was motivated solely by the need to have a floating autosegment, the ~e. It is equivalent to collapse these two tiers, using the alternation ce~0 in place of the floating ce. This style of approach to zero alternations, which dates back to Bloomfield (1926), will employ the parenthesis notation for optional items that was defined in Section 2.1. We follow Tranel in representing syllable structure and we shall do this using the notation shown in (52). 16 16 Our analysis is not crucially tied to this particular version of syllable structure, which is most closely related to the proposals of Kahn (1976) and Clements and Keyser (1983).</Paragraph>
      <Paragraph position="3">  An independent tier that represents syllable structure will be encoded as a sequence of such syllables, where the segmental constituents of the syllable structure are coindexed with a separate segmental tier, as defined in (53). Note that the indices in (53) range over lists that may be empty in the case of onsets and codas, and that the type phrase denotes phonological phrases.</Paragraph>
      <Paragraph position="4">  The notation of (53) states that in order for something to be a well-formed phrase, its sequence of segments must be parsed into a sequence of well-formed syllables. In more familiar terms, one could paraphrase (53) as stating that the domain of syllabification in French is the phrase.</Paragraph>
      <Paragraph position="5"> As a simple illustration of the approach, consider again the word melons. The proposed lexical representation for the phonology attribute of this word is \[SEGS : (m (ce) 1 5}\]. When we insist that any phrase containing this word must consist of a sequence of well-formed syllables, we can observe the following pattern of behavior for six melons.</Paragraph>
      <Paragraph position="6">  Observe in the above example that the syllabic position of m is variable. In Example 54a m is in an onset while in 54b it is in a coda. Therefore, it is inappropriate to  Parts of French type hierarchy.</Paragraph>
      <Paragraph position="7"> insist that the syllabic affiliation of segments is determined lexically. Rather, we have opted for the prosodic type phrase, insisting that anything of this type consists of one or more well-formed syllables (cf. Example 11).</Paragraph>
      <Paragraph position="8"> Now consider the case of the phrase sept melons. This is similar to the situation in (54), except that we must find a way of ruling out the tml cluster as a valid coda-onset sequence. We are not aware of any exhaustive study of possible French consonant clusters, although one can find discussions of particular clusters (e.g., Tranel \[1987b:95ff\] shows that CLj onset clusters are not tolerated). Consequently, the two hierarchies in Figure 3 are necessarily preliminary, and are made more for the sake of being explicit than for their precise content. Note that parentheses indicate optionality, so, for example, both onsets and codas are allowed to be null. Additional stipulations will be necessary to ensure that an intervocalic consonant is syllabified with the material to its right. We can do this by preventing an onsetless syllable from following a closed syllable, with the type onset-max-1.</Paragraph>
      <Paragraph position="9">  Observe that this list of syllables contains a violation of (55), so \[si.moel.5\] is ruled out. Now that we have considered vowel-consonant-vowel (VCV) sequences, we shall move on to more complex intervocalic consonant clusters.</Paragraph>
      <Paragraph position="10"> Although the constraints in Figure 3 produce the desired result for VLLV clusters (L--liquid), by assigning each liquid to a separate syllable (Tranel 1987b), there is still ambiguity with VOLV clusters (O=obstruent), which are syllabified as V.OLV according to Tranel. We can deal with this and similar ambiguities by further refining  This constraint states that it is not permissible to have an obstruent in a syllable coda if the following onset lacks an obstruent. Equivalently, we could say that if a syllable coda contains an obstruent then the following onset must also contain an obstruent. To see why these constraints are relevant to schwa, consider the case of demanderions, (also discussed by Tranel \[1987b\]). The constraints in Figure 3 rule out * \[doe.m~t.drjS\], since the underlined onset cluster is too complex. The constraint in (57) rules out *\[doe.m0~d.rj6\], where the obstruent d is assigned to the preceding syllable to leave an rj onset. The remaining two possible pronunciations are \[dce.m~t.dc~.rj6\] and \[dce.m0~.dri.j6\], as required. (Note that the ions suffix has the two forms, \[j6\] and \[ij6\].) Now let us consider the case of h-aspir6 words. These vowel-initial words do not tolerate a preceding consonant being syllabified into the word-initial onset. What happens to the V.CV and V.OLV constraints when the second vowel is in the first syllable of an h-aspir6 word, as we find in sept haches \[s~t.a ~\], *\[s~.ta~\] and quatre haches \[katr.ay\], *\[kat.ray\], *\[ka.tray\]? Here, it would appear that Tranel's analysis breaks down. Our conjecture is that the constraints in (55) and (57) should only apply when the second syllable is not an h-aspir6 syllable. So we need to introduce a further distinction in syllable types, introducing ha-syl for h-aspir6 syllables and nha-syl for the rest.</Paragraph>
      <Paragraph position="11">  ha-syl I.</Paragraph>
      <Paragraph position="12"> Accordingly, the constraints (55) and (57) are refined, so that the second syllable is of the type nha-syl. The revised constraints are given in (60).</Paragraph>
      <Paragraph position="13"> Example 60 &lt; CODa ne,/st a. onset-max-l' -- -~ &amp;quot; syl nha-syl phrase b. &lt; ICOD :&lt;o &gt;\] \[ONS: &lt;o s &gt;\]&gt; onset-max-2' =-- ~ &amp;quot;&amp;quot; syl nha-syl phrase Now, h-aspir6 words will be lexically specified as having an initial ha-syl. However, we must not specify any more syllable structure than is absolutely necessary. Example 61 displays the required constraint for the word haut.</Paragraph>
      <Paragraph position="14">  So although syllabification operates at the phrase level rather than the morpheme level (see Example 53), we are still able to impose lexically conditioned constraints on syllable structure directly.</Paragraph>
      <Paragraph position="15"> It remains to be shown how this treatment of h-aspir6 bears on schwa. Fortunately, Tranel (1987b:94) has provided the example we need. Consider the phrase dans le haut \[d~t.lce.o\]. This contains the word le \[l(ee)\], which is lexically specified as having an optional ce, indicated by parentheses. 17 There are three possible syllabifications, only the last of which is well formed.</Paragraph>
      <Paragraph position="16">  The syllabification in (62a) is unavailable, since the syllable corresponding to the word haut is lexically specified as ha-syl, which means that its onset must be an elist from (59). The syllabifications in (62b) are likewise unavailable since these both consist of a syllable with a coda followed by a syllable without an onset, in contravention of (60a). This only leaves (62c), which corresponds to the attested form \[d~t.lce.o\].</Paragraph>
      <Paragraph position="17"> We conclude this section with an example derivation for the phrase on ne se moque pas \[6n.smak.pct\], which was presented in (51). We assume that at some stage of a derivation, the PHON attribute of a sign is as follows:  Computational Linguistics Volume 20, Number 3 When the appropriate grammatical conditions are met, this phonology attribute will be given the type phrase. The definition in (53) will accordingly specialize the SYLS attribute. One possible specialization :is given in Example 64.</Paragraph>
      <Paragraph position="18"> Example 64 f /\[ degNs: rONBET:,BI,1 YLS: NUC: I NUC: / CODA: LCODA: {k) J syl syl phrase LSEGS : (6 n s m o k p ct) (P)1\1 ,NUC. j;\] Lcoo -I / The reader can check that the onset and coda sequences comply with the constraints in Figure 3, that the first syllable can have an empty onset because there is no preceding syllable that could have a coda that matches the requirements of (60a), and that the obstruent k is permitted by constraint (60b) to appear in the coda of the second syllable because there is another obstruent p in the following onset.</Paragraph>
      <Paragraph position="19"> This concludes our discussion of French schwa. We believe our treatment of schwa is empirically equivalent to that of Tranel (1987a), except for the analysis of h-aspir6. Several empirical issues remain, but we are optimistic that further refinements to our proposals will be able to take additional observations on board. Notwithstanding such further developments, we hope to have demonstrated that the procedural devices of deletion and rule ordering are unnecessary in a typed feature-based grammar framework, and that constraints represent a perspicuous way of encoding linguistic observations.</Paragraph>
    </Section>
  </Section>
  <Section position="7" start_page="485" end_page="487" type="metho">
    <SectionTitle>
6. Prospects for Implementation
</SectionTitle>
    <Paragraph position="0"> In the preceding sections we have shown how the use of parameterized lists in HPSG is sufficient for encoding a variety of phonological generalizations. While we like this approach for the purposes of specification and exposition, as stated in Section 1.4, we actually envisage an implementation employing finite-state automata for string manipulation. This is simply because we favor the use of existing well-understood technology when it comes to producing an efficient implementation.</Paragraph>
    <Paragraph position="1"> As we have already explained in Section 1.4, we have linguistic reasons for not wishing to use finite-state transducers and the concomitant two-level model, and instead are interested in exploring the prospects of integrating our work with the automaton model of Bird and Ellison (1994). In this section we give an overview of this automaton model and briefly outline the view of automata as types that was originally proposed in Bird (1992).</Paragraph>
    <Section position="1" start_page="485" end_page="487" type="sub_section">
      <SectionTitle>
6.1 One-Level Phonology
</SectionTitle>
      <Paragraph position="0"> For a variety of reasons already laid out in Section 1, we would like to achieve a closer integration between phonology and constraint-based grammar frameworks like HPSG.</Paragraph>
      <Paragraph position="1"> However, for such an integration to work, it is necessary to adopt a rather unusual view of phonology; one characterized by such notions as compositionality, intensionality, and lexicalism, and which has come to be called constraint-based phonology (Bird 1990).</Paragraph>
      <Paragraph position="2"> Recently, Bird and Ellison (1994) have reinterpreted the constraint-based approach to phonology using finite-state automata. Nonlinear phonological representations and  Two views of autosegmental association.</Paragraph>
      <Paragraph position="3"> rules are encoded as automata. The key insight is that if autosegmental association is viewed as overlap between intervals with duration (Bird and Klein 1990), then the overlap can be simulated by using synchronization primitives on automata. Figure 4 illustrates this idea. The diagram on the left of Figure 4 shows two temporal intervals x and y that overlap during the shaded period. On the right, the intervals x and y themselves are represented as sequences of contiguous tape cells where each cell contains a copy of the appropriate information (here, simply repeats of x and y). Again, the shaded period indicates the period of 'overlap' of the two intervals. The reader is referred to Bird and Ellison (1994) for further details.</Paragraph>
      <Paragraph position="4"> Although this kind of phonology employs formal devices very similar to the two-level FST model, there are some important differences in how the two models are used. In the two-level model the traditional distinction in phonology between RULES and REPRESENTATIONS is evident in the transducers and tapes respectively. As in constraint-based grammar more generally, one-level phonology does not have this distinction; rules and representations alike are interpreted as automata. Figure 5 illustrates this difference.</Paragraph>
      <Paragraph position="5"> Now that we have outlined the one-level model and briefly discussed its relationship with the two-level model, we shall sketch the link to typed feature systems.  Computational Linguistics Volume 20, Number 3</Paragraph>
    </Section>
    <Section position="2" start_page="487" end_page="487" type="sub_section">
      <SectionTitle>
6.2 Types as Automata
</SectionTitle>
      <Paragraph position="0"> A type denotes a set of objects. Thus, types are descriptions, and they can be combined using the familiar operations of meet, join, and negation. Similarly, an automaton denotes a set of objects, namely strings (o1&amp;quot; automaton tapes). And likewise, the operations of meet, join, and negation are defined for automata and correspond to intersection, union, and complement of the corresponding sets. Of course, a further operation of concatenation is defined for automata. We envisage a system for processing linguistic descriptions that implements a subset of the types (which we might simply call STRING types) as finite-state automata over some predefined alphabet. When the inference engine requires that two string types be 'unified,' the meet of the corresponding automata will be formed.</Paragraph>
      <Paragraph position="1"> Although these string types may be declared as the appropriate values for certain attributes in a typed feature system, string types are only declared in terms of the basic alphabet and other string types. It is not possible to employ non-string types in the definition of string types. This is a severe restriction, since list types (say, in HPSG) allow arbitrary feature structures as elements, and we would like to be able to do the same for string types. Work on overcoming this limitation is currently in progress, and builds on the well-known similarity between feature structures and automata, when viewed as directed graphs (Kasper and Rounds 1986).</Paragraph>
    </Section>
  </Section>
class="xml-element"></Paper>
Download Original XML