File Information
File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/abstr/00/a00-2017_abstr.xml
Size: 1,445 bytes
Last Modified: 2025-10-06 13:41:33
<?xml version="1.0" standalone="yes"?> <Paper uid="A00-2017"> <Title>A Classification Approach to Word Prediction*</Title> <Section position="1" start_page="0" end_page="0" type="abstr"> <SectionTitle> Abstract </SectionTitle> <Paragraph position="0"> The eventual goal of a language model is to accurately predict the value of a missing word given its context. We present an approach to word prediction that is based on learning a representation for each word as a function of words and linguistics predicates in its context. This approach raises a few new questions that we address. First, in order to learn good word representations it is necessary to use an expressive representation of the context. We present a way that uses external knowledge to generate expressive context representations, along with a learning method capable of handling the large number of features generated this way that can, potentially, contribute to each prediction. Second, since the number of words &quot;competing&quot; for each prediction is large, there is a need to &quot;focus the attention&quot; on a smaller subset of these. We exhibit the contribution of a &quot;focus of attention&quot; mechanism to the performance of the word predictor. Finally, we describe a large scale experimental study in which the approach presented is shown to yield significant improvements in word prediction tasks.</Paragraph> </Section> class="xml-element"></Paper>