File Information

File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/intro/06/p06-2119_intro.xml

Size: 5,809 bytes

Last Modified: 2025-10-06 14:03:46

<?xml version="1.0" standalone="yes"?>
<Paper uid="P06-2119">
  <Title>Word Sense Disambiguation using lexical cohesion in the context</Title>
  <Section position="3" start_page="1" end_page="929" type="intro">
    <SectionTitle>
1 Introduction
Word Sense Disambiguation (WSD) is generally
</SectionTitle>
    <Paragraph position="0"> taken as an intermediate task like part-of-speech (POS) tagging in natural language processing, but it has not so far achieved the sufficient precision for application as POS tagging (for the history of WSD, cf. Ide and Veronis (1998)). It is partly due to the nature of its complexity and difficulty, and to the widespread disagreement and controversy on its necessity in language engineering, and to the representation of the senses of words, as well as to the validity of its evaluation (Kilgarriff and Palmer, 2000). However the endeavour to automatically achieve WSD has been continuous since the earliest work of the  2000) as these are well known and accepted in the community of computational linguistics.</Paragraph>
    <Paragraph position="1"> With respect to whether or not they employ the training materials provided, SENSEVAL roughly categorizes the participating systems into &amp;quot;unsupervised systems&amp;quot; and &amp;quot;supervised systems&amp;quot;. Those that don't use the training data are not usually truly unsupervised, being based on lexical knowledge bases such as dictionaries, thesauri or semantic nets to discriminate word senses; conversely the &amp;quot;supervised&amp;quot; systems learn from corpora marked up with word senses.</Paragraph>
    <Paragraph position="2"> The fundamental assumption, in our &amp;quot;unsupervised&amp;quot; technique for WSD in this paper, is that the similarity of contextual features of the target with the pre-defined features of its sense in the lexical knowledge base provides a quantitative cue for identifying the true sense of the target. null The lexical ambiguity of polysemy and homonymy, whose distinction is however not absolute as sometimes the senses of word may be intermediate, is the main object of WSD. Verbs, with their more flexible roles in a sentence, tend to be more polysemous than nouns, so worsening the computational feasibility. In this paper we disambiguated the sense of a word after its POS tagging has assigned them either a noun or a verb tag. Furthermore, we deal with nouns and verbs separately.</Paragraph>
    <Paragraph position="3"> 2 Some previous work on WSD using semantic similarity Sussna (1993) utilized the semantic network of nouns in WordNet to disambiguate term senses to improve the precision of SMART information retrieval at the stage of indexing, in which he assigned two different weights for both directions of edges in the network to compute the similarity of two nodes. He then exploited the moving fixed size window to minimize the sum  of all combinations of the shortest distances among target and context words.</Paragraph>
    <Paragraph position="4"> Pedersen et al. (2003) extended Lesk's definition method (1986) to discriminate word sense through the definitions of both target and its IS-A relatives, and achieved a better result in the English lexical sample task of SENSEVAL-2, compared with other edge-counting or statistical estimation metrics on WordNet.</Paragraph>
    <Paragraph position="5"> Humans carefully select words in a sentence to express harmony or cohesion in order to ease the ambiguity of the sentence. Halliday and Hasan (1976) argued that cohesive chains unite text structure together through reiteration of reference and lexical semantic relations (superordinate and subordinate). Morris and Hirst (1991) suggested building lexical chains is important in the resolution of lexical ambiguity and the determination of coherence and discourse structure. They argued that lexical chains, which cover the multiple semantic relations (systematic and nonsystematic), can transform the context setting into the computational one to narrow down the specific meaning of the target, manually realizing this with the help of Roget's Thesaurus. They defined a lexical chain within Roget's very general hierarchy, in which lexical relationships are traced through a common category.</Paragraph>
    <Paragraph position="6"> Hirst and St-Onge (1997) define a lexical chain using the syn/antonym and hyper/hyponym links of WordNet to detect and correct malapropisms in context, in which they specified three different weights from extra-strong to medium strong to score word similarity to decide the inserting sequence in the lexical chain. They first computationally employed WordNet to form a &amp;quot;greedy&amp;quot; lexical chain as a substitute of the context to solve the matter of malapropism, where the word sense is decided by its preceding words.</Paragraph>
    <Paragraph position="7"> Around the same time, Barzilay and Elhadad (1997) realized a &amp;quot;non-greedy&amp;quot; lexical chain, which determined the word sense after processing of all words, in the context of text summarization. null In this paper we propose an improved lexical chain, the lexical hub, that holds the target to be disambiguated as the centre, replacing the usual chain topology used in text summarization and cohesion analysis. In contrast with previous methods we only record the lexical hub of each sense of the target, and we don't keep track of other context words. In other words, after the computation of lexical hub of the target, we can immediately produce the right sense of the target even though the senses of the context words are still in question. We also transform the context surroundings through a word association thesaurus to explore the effect of other semantic relationships such as syntagmatic relation against WSD.</Paragraph>
  </Section>
class="xml-element"></Paper>
Download Original XML