File Information
File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/88/j88-3014_metho.xml
Size: 10,860 bytes
Last Modified: 2025-10-06 14:12:14
<?xml version="1.0" standalone="yes"?> <Paper uid="J88-3014"> <Title>DISTINGUISHING USER MODELS FROM DISCOURSE MODELS</Title> <Section position="3" start_page="0" end_page="0" type="metho"> <SectionTitle> 2 DEFINING USER MODELS AND DISCOURSE MODPLS </SectionTitle> <Paragraph position="0"> Some authors define user models simply as information that the system has about its users (e.g., Schuster). I think this definition is too broad. Consider an NL interface to a data base, which contains the following relation: 26 Jones 32 40 When Mr. Jones happens to be the user of this system and asks, &quot;What is my bonus?&quot;, the system should respond &quot;40&quot;. In this case, the system has information about the user, but one would not like to say that its response was based on a user model.</Paragraph> <Paragraph position="1"> Even if one restricts the definition above to &quot;information about the user put to use&quot; (see Sparck Jones), it is not strong enough. If a deductive data base in addition to the relation above includes a rule like &quot;If AGE(X) > 30 and BONUS(X) > 35 then STATUS(X) = 10&quot; and Mr. Jones asks, 'degWhat is my status?&quot; the system should respond &quot;10&quot;. Even though the deductive DBMS uses information about the user to instantiate the inference rule, such a system should not be construed as having a user model.</Paragraph> <Paragraph position="2"> I propose the following joint definitions of user model and user modeling component (see Wahlster and Kobsa 1988) as well as discourse model and discourse modeling component in the context of NL dialog systems: A user model is a knowledge source that contains explicit assumptions on all aspects of the user that may be relevant for the dialog behavior of the system. A user modeling component is that part of a dialog system whose function is to tions about the user.</Paragraph> <Paragraph position="3"> A discourse model is a knowledge source that contains the system's description of the syntax, semantics, and pragmatics of a dialog as it proceeds. A discourse modeling component is that part of a dialog system whose function is to * incrementally build up a discourse model; * store and update entries in it; and * supply other components of the system with information about the structure and content of the previous segments of the dialog.</Paragraph> <Paragraph position="4"> While it seems commonly agreed upon that a DM should contain a syntactic and semantic description of discourse segments, a record of the discourse entities mentioned, the attentional structure of the dialog including a focus space stack, anaphoric links, and descriptions of individual utterances on the speech act level, there seem to be many other ingredients needed for a good discourse representation which are not yet worked out in current computational discourse theory. Therefore, I prefer to refer only to the abstract levels of necessary discourse representation in the definition above.</Paragraph> </Section> <Section position="4" start_page="0" end_page="0" type="metho"> <SectionTitle> 3 SOME DIFFERENCES AND SIMILARITIES BETWEEN USER MODELS AND DISCOURSE MODELS </SectionTitle> <Paragraph position="0"> An important difference between a discourse model and a user model is that entries in the user model often must be explicitly deleted or updated, whereas in the discourse model entries describing the structure and content of utterances of the ongoing dialog are never deleted (except for forgetting phenomena, which are beyond the scope of the current discussion). Thus, according to our definition above, a belief revision component is an important part of a user modeling component.</Paragraph> <Paragraph position="1"> Consider the following dialog with a hypothetical tutoring system in the SCHOLAR tradition.</Paragraph> <Paragraph position="2"> System: (1) Tell me about California.</Paragraph> <Paragraph position="3"> User: (2) San Francisco is the capital of California.</Paragraph> <Paragraph position="4"> System: (3) No, that's wrong.</Paragraph> <Paragraph position="5"> User: (4) I see. So, that's not the capital.</Paragraph> <Paragraph position="6"> (5) Then, what is its capital? System: (6) Sacramento.</Paragraph> <Paragraph position="7"> (7) Now, tell me why you mentioned San Francisco first, when you began to talk about California.</Paragraph> <Paragraph position="8"> A simple consequence of the user's response (2) is an entry in the system's user model, which represents the fact, that the system believes that the user believes (B1). After (3), and certainly after (4) the user model should contain (BI').</Paragraph> <Paragraph position="10"> This means that the user modeling component has to remove (B1) from the user model (in a reason maintenance system this causes (B1) to be added to the set of beliefs, which are currently &quot;out&quot;). After (6) the user's belief (B2) should be added to the system's user model.</Paragraph> <Paragraph position="11"> If the apriori user model contains &quot;For each state there exists one and only one capital&quot; as a mutual believed fact, then the user modeling component can also remove (BI') after adding (B2).</Paragraph> <Paragraph position="12"> In the discourse model, of course, the fact that the user uttered sentence (2) should not be deleted. For example, the system could go on and ask the user a question like (7), which explicitly refers to the fact that (2) was the first reaction to (I). What this simply means is that the fact that the user made a particular assertion 102 Computational Linguistics, Volume 14, Number 3, September 1988 Wolfgang Wahlster Distinguishing User Models from Discourse Models remains true even if the user's belief changes and he withdraws his previous assertion.</Paragraph> <Paragraph position="13"> Even a metacommunicative act like (9) should not delete entries in the discourse model, as the successful anaphoric reference in (10) to a discourse entity introduced in (8) suggests. But it is obvious that in the user model the corresponding representation of the user's wants has to be changed.</Paragraph> <Paragraph position="14"> User:(8) I don't want to travel with my kids.</Paragraph> <Paragraph position="15"> (9) Forget what I just said.</Paragraph> <Paragraph position="16"> (10) I want to travel with them.</Paragraph> <Paragraph position="17"> This does not imply that the discourse model is static and the user model is dynamic. The discourse model is also highly dynamic (consider, e.g., focus shifting), but it lacks the notion of logical consistency, which is important for belief revision and default reasoning in a user modeling component. In my view, the discourse model is like an annotated trace of the various levels of the system's processing involved in understanding the user's utterances and generating its own dialog contributions. null Let's consider another example to emphasize the differences between a UM and a DM. Suppose that the system plays the role of a travel agent, who wants to sell trips to the well-known holiday places A and B, for which it has some reasonably priced offers. When the user asks, &quot;What are your cheapest trips?&quot; the system lists A and B first, followed by a hastily presented list of eight other places with names, which it assumes are totally unfamiliar to the user. In the system's DM all ten places appear, but the user modeling component of the system explicitly assumes that the user only believes &quot;cheap-trip-to(A)&quot;, &quot;cheap-trip-to(B)&quot; together with the belief that there are some other cheap trips available. This is exactly the aim of the uncooperative behavior of the travel agent: Now, it is likely that the user wants to know more about the offers A and B, which the agent wants to sell. But if the user later finds out that a trip to one of the other places is much cheaper and better, and complains to the travel agent, &quot;Why didn't you suggest this trip right at the beginning?&quot;, the travel agent can refer back to his DM and say, &quot;I mentioned this place among my first suggestions&quot;.</Paragraph> <Paragraph position="18"> Some authors claim that the discourse model expires at the end of a dialog, while parts of the user model may be saved for further use (e.g., Chin). I think that is wrong. Often a dialog participant is able to paraphrase a segment of a previous dialog without remembering who the dialog partner was and at what time and location the dialog took place. While he may not be able to reconstruct the exact phrasing, he has access to a representation of the semantics and pragmatics of the interaction. Furthermore, I think that often conversational rules and tactics are learned by induction over a large set of interaction patterns extracted from discourse models, which were partially saved in episodic memory, where they are not necessarily associated with a long-term user model. In order to learn how to use a language it seems to be important to not always discard the complete discourse model after the end of a conversation. null On the other hand, one often has many assumptions about the beliefs, plans, and goals of a dialog partner before a new dialog begins (cf. Wahlster and Kobsa 1988), without having a clear idea from which actual dialogs the assumptions in this user model were derived. Thus I agree with Morik that the short-term/ long-term criterion cannot be used to distinguish user models and discourse models. If one prefers to restrict the term discourse model to an ongoing conversation and to define the saved portions of it as part of the world knowledge, then one should do the same for the term user model, so that again the criterion does not discriminate. null While in many cases the UM component and the DM component process the same input (e.g., a meaning representation of the last utterance), and their output is used by the same processes, I would suggest that both components be kept separate. Even if there is some information, which should be present in both models, it will be represented in another form, since, as I pointed out above, the functionality and the type of processing of the UM and DM components are so different. In this case we have a multiple knowledge representation in the UM and the DM, which is quite common in complex A1 systems.</Paragraph> <Paragraph position="19"> As I remarked in the beginning, in asking who is right in this discussion, one must carefully evaluate the corresponding definitions for UM and DM proposed by the respective authors. In this paper, I introduced and motivated definitions, under which a UM and a DM are separate, but related to each other.</Paragraph> </Section> class="xml-element"></Paper>