File Information
File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/metho/05/w05-0208_metho.xml
Size: 22,612 bytes
Last Modified: 2025-10-06 14:09:52
<?xml version="1.0" standalone="yes"?> <Paper uid="W05-0208"> <Title>Towards a Prototyping Tool for Behavior Oriented Authoring of Conversational Agents for Educational Applications</Title> <Section position="5" start_page="46" end_page="46" type="metho"> <SectionTitle> 3 Preliminary Design Intents from Contextual Interviews </SectionTitle> <Paragraph position="0"> The core essence of the user-centered design approach is designing from data rather than from preconceived notions of what will be useful and what will work well. Expert blind spots often lead to designs based on intuitions that overlook needs or overly emphasize issues that are not centrally important (Koedinger & Nathan, 2004; Nathan & Koedinger, 2000). Contextual inquiry is used at an early stage in the user-centered design process to collect the foundational data on which to build a design (Beyer and Holtzbatt, 2000). Contextual Inquiry is a popular method developed within the Human Computer Interaction community where the design team gathers data from end users while watching what the users do in context of their work. Contextual interviews are used to illuminate these observations by engaging end-users in interviews in which they show specific instances within their work life that are relevant for the design process.</Paragraph> <Paragraph position="1"> These methods help define requirements as well as plan and prioritize important aspects of functionality. At the same time, the system designers get a chance to gain insights about the users' environment, tasks, cultural influences and difficulties in the current processes.</Paragraph> <Paragraph position="2"> Many aspects of the Tutalk tool were designed based on contextual inquiry (CI) data.</Paragraph> <Paragraph position="3"> The design team conducted five CIs with users who have experience in using existing authoring tools such as Carmel-Tools (Rose & Hall, 2004).</Paragraph> <Paragraph position="4"> The design team leader also spent one week observing novice tool users working with the current set of tools at an Intelligent Tutoring Summer School. Here we will discuss some findings from those CIs and observations and how they motivated some general design intents, which we flesh out later in the paper.</Paragraph> <Paragraph position="5"> A common pattern we observed in our CIs was that having different floating windows for different tasks fills up the computer screen relatively quickly and confuses authors as to where they are in the process of authoring. The TuTalk design addresses this observed problem by anchoring the main window and switching only the components of the window as needed. A standard logic for layout and view switching helps authors know what to expect in different contexts. Placement of buttons in TuTalk is consistently near the textboxes that they control, and a bounding box is drawn around related sets of controls so that the user does not get lost trying to figure out where the buttons are or what they are for.</Paragraph> <Paragraph position="6"> We observed that authors needed to refer to cheat sheets and user documentation to use their current tools effectively and that different users did not employ the same terminology to refer to similar functionality, which made communication difficult. Furthermore, their current suites of tools were not designed as one integrated environment. Thus, a lot of shuffling of files from one directory to another was required in order to complete the authoring process. Users without Unix operating system experience found this especially confusing. Our goal is to require only very minimal documentation that can be obtained on-line in the context of use.</Paragraph> <Paragraph position="7"> TuTalk is a single, integrated environment that makes use of GUI widgets for actions rather then requiring any text-based commands or file system activity. In this way we hope to avoid requiring the users to use a manual or a &quot;cheatsheet&quot; reference for the commands they forget. As is common practice, TuTalk also uses consistent labels throughout the interface to promote understandability and communication with tool developers as well as other dialogue system developers. null</Paragraph> </Section> <Section position="6" start_page="46" end_page="47" type="metho"> <SectionTitle> 4 Exploring the User's Mental Model </SectionTitle> <Paragraph position="0"> through User Studies As an additional way of gaining insights into what sort of interface would make the process of authoring conversational interfaces accessible, we conducted a small, exploratory user study in which we examined how members of our target user population think about the structure of language. null Two groups of college-level participants with no deep linguistics training were asked to read three transcribed conversations about ordering from a menu at a restaurant from our English as a Second Language corpus. The three specific restaurant dialogues were chosen because of their breadth of topic coverage and richness in linguistic expression. Participants were asked to perform tasks with these dialogues to mimic three levels of conversational interface authoring: null Macro Organization Tasks (dialogue level) Level 1. How authors understand, segment, and organize dialogue topics Level 2. How authors generalize across dialogues as part of constructing a &quot;model&quot; script Micro Organization Task (sentence level) Level 3. How authors categorize and decompose sentences within these dialogues null The first group (Group A, five participants) was asked to perform Macro Organization Tasks before processing sentences for the Micro Organization Tasks. The second group (Group B, four participants) was asked to perform these sets of tasks in the opposite order.</Paragraph> <Paragraph position="1"> Our findings for the Macro Organization Tasks showed that participants effectively broke down dialogues into segments that reflected intuitive breaks in the conversation. These topics were then organized into semantically related categories. Although participants were not explicitly instructed on how to organize the topics, every participant used spatial proximity as a representation for semantic relatedness. Another finding was the presence of primacy effects in the &quot;model&quot; restaurant scripts they were asked to construct. These scripts were heavily influenced by the first dialogue read. As a result, important topics that surfaced in the other two dialogues were omitted from the model scripts.</Paragraph> <Paragraph position="2"> Furthermore, we found that participants in Group B took much longer in completing the Micro Organization Task (35-40 minutes as opposed to 25-30 minutes) without performing the Macro Organization Tasks first. In general, we found that participants clustered sentences based on surface characteristics rather than creating ontologically similar classes that would be more useful from a system development perspective.</Paragraph> <Paragraph position="3"> In a follow-up study we are exploring ways of guiding users to cluster sentences in ways that are more useful from a system building perspective. null Our preliminary findings show that getting an overall sense of the corpus facilitates micro-level organization. This is hindered by two factors: First, primacy effects interfere with macro-level comprehension. Second, system developers struggle to strategically select portions of their corpus on which to focus their initial efforts.</Paragraph> </Section> <Section position="7" start_page="47" end_page="49" type="metho"> <SectionTitle> 5 Stage One: Corpus Organization </SectionTitle> <Paragraph position="0"> While existing tools from our previous work required authors to organize their corpus data prior to their interaction with the tools, both our contextual research and user studies indicated that support for organizing corpus data prior to authoring is important.</Paragraph> <Paragraph position="1"> In light of this concern, the TuTalk authoring process consists of three main stages. Corpus collection, corpus data organization through what we call the InfoMagnet interface, and authoring propper. First, a corpus is collected by asking users to engage in conversation using either a typed or spoken chat interface. In the case of spoken input, the speech is then transcribed into textual form. Second, the raw corpus data is automatically preprocessed for display and interactive organization using the InfoMagnet interface. As part of the preprocessing, dialogue protocols are segmented automatically at topic boundaries, which can be adjusted by hand later during authoring propper. The topic oriented segments are then clustered semi-automatically into topic based classes. The output from this stage is an XML file where dialogue segments are reassembled into their original dialogue contexts, with each utterance labeled by topic. This XML file is finally passed onto the authoring environment propper, which is then used for finer grained processing, such as shifting topic segment boundaries and labeling more detailed utterance functionality.</Paragraph> <Paragraph position="2"> Our design is for knowledge sources that are runable from our dialogue system engine to be generated directly from the knowledge base created during the fine-grained authoring process as in Carmel-Tools (Rose & Hall, 2004), however currently our focus is on iterative development of a prototype of the authoring interaction design. Thus, more work is required to create the final end-to-end implementation. In this section we focus on the design of the corpus collection and organization part of the authoring process.</Paragraph> <Section position="1" start_page="48" end_page="48" type="sub_section"> <SectionTitle> 5.1 Corpus Collection </SectionTitle> <Paragraph position="0"> An important part of our mission is developing technology that can use collected and automatically pre-processed corpus data to guide and streamline the authoring process. Prior to the arduous process of organizing and extracting meaningful data, a corpus must be collected.</Paragraph> <Paragraph position="1"> As part of the PSLC and other local tutorial dialogue efforts we have collected corpus data from multiple domains that we have made use of in our development process. In particular, we have been working with data collected in connection with the PSLC Physics and English as a Second Language LearnLab courses as well as local Calculus and Thermodynamics tutoring projects. Currently we have physics tutoring data primarily from one physics tutor (interactions with 40 students), thermodynamics data from four different tutors (interactions with 27 students), Calculus data from four different tutors (84 dialogues), and ESL dialogues collected from 15 pairs of students (30 dialogues altogether). null While we have drawn upon data from all of these domains for testing the underlying language processing technology for our development effort, for our user studies we have so far mainly drawn upon our ESL corpus, which includes conversations between students about every-day tasks such as ordering from a restaurant or about their pets. We chose the language ESL data for our initial user tests because we expected it to be easy for a general population to relate to, but we plan to begin using calculus data as well.</Paragraph> </Section> <Section position="2" start_page="48" end_page="49" type="sub_section"> <SectionTitle> 5.2 InfoMagnets Interface </SectionTitle> <Paragraph position="0"> As mentioned previously, once the raw dialogue corpus is collected, the next step is to sift through this data and assign utterances (or groups of utterances) to classes conceptualized by the author. Clustering is a natural step in this kind of exploratory data analysis, as it promotes learning by grouping and generalizing from what we know about some of the objects in a cluster. For this purpose we have designed the InfoMagnets interface, which introduces a non-technical metaphor to the task of iterative document clustering. The InfoMagnets interface was designed to address the problems identified in the user study discussed above in Section 4.</Paragraph> <Paragraph position="1"> Specifically, we expected that those problems could be addressed with an interface that: 1. Divides dialogues into topic based segments and automatically clusters them into conceptually similar classes 2. Eliminates primacy effects of sequential dialogue consumption by creating an inclusive compilation of all dialogue topics 3. Makes the topic similarity of docu- null ments easily accessible to the user The InfoMagnets interface is displayed in respond to cluster centroids and the smaller ones (particles) correspond to actual spans of text. Lexical cohesion in the vector space translates into attraction in the InfoMagnet space. The attraction from each particle to each InfoMagnet is evident from the particle's position with respect to all InfoMagnets and its reaction-time when an InfoMagnet is moved by the user, which causes the documents that have some attraction with it to redistribute themselves in the InfoMagnet space.</Paragraph> <Paragraph position="2"> Figure 1 InfoMagnets Interface Being an unsupervised learning method, clustering often requires human-intervention for fine-tuning (e.g. removing semantically-weak discriminators, culling meaningless clusters, or deleting/splitting clusters too fine/coarse for the author's purpose). The InfoMagnets interface provides all this functionality, while shielding the author from the computational details inherent in these tasks Initially, the corpus is clustered using the Bisecting K-means Algorithm described in (Kumar et al., 1998). Although this is a hard clustering algorithm, the InfoMagnet interface shows the particles association with all clusters, given by the position of the particle. Using a cross-hair lens, the author is able to view the contents of each cluster centroid and each particle. The author is able to select a group of particles and view the common features between these particles and any InfoMagnet in the space. The interface allows the editing of InfoMagnets by adding and removing features, splitting InfoMagnets, and removing InfoMagnets. When the user edits an InfoMagnets, the effect in the particle distribution is shown immediately and in an animated way.</Paragraph> </Section> <Section position="3" start_page="49" end_page="49" type="sub_section"> <SectionTitle> 5.3 XML format </SectionTitle> <Paragraph position="0"> The data collected from the conversations in .txt format are reformatted into XML format before being displayed with InfoMagnet tool.</Paragraph> <Paragraph position="1"> The basic XML file contains a transcription of the conversational data and has the following structure: Under the top root tag, there is <dialogue> tag which designates the conversion about a topic. It has an &quot;id&quot; attribute so that we can keep track of each separate conversation.</Paragraph> <Paragraph position="2"> Then each sentence has a <sentence> tag with two attributes &quot;uid&quot; and &quot;agent&quot;. &quot;uid&quot; is a universal id and &quot;agent&quot; tells who was speaking. Additionally, sentences are grouped into segments, marked off with a <subtopic> tag.</Paragraph> <Paragraph position="3"> The user's interaction with the InfoMagnet interface adds a &quot;subtopic-name&quot; attribute to the subtopic tag. Then, the authoring interface proper, described below, allows for further adjustments and additions to the xml tags. The final knowledge sources will be generated from this XML based representation.</Paragraph> </Section> </Section> <Section position="8" start_page="49" end_page="51" type="metho"> <SectionTitle> 6 Authoring </SectionTitle> <Paragraph position="0"> The authoring environment proper consists of two main views, namely the authoring view and tutoring view. The authoring view is where the author designs the behavior of the conversational agent. The authoring view has two levels; the topic level and the subtopic level. The tutoring view is what a student will be looking at when interacting with the conversational agent.</Paragraph> <Paragraph position="1"> Our focus here is on the Authoring view.</Paragraph> <Paragraph position="2"> Authoring View: Topic Level The Topic level of the authoring view allows for manipulating the relationship between subtopics as well as the definition of the subtopic. Figure 2 shows the topic level authoring view, which consists of two panels. In the left, the author inputs the description of the task that the student will engage in with the agent. The author can specify whether the student will be typing or talking, the title of the topic, the task description, an optional picture that aids with the task (such as a menu or a map of a city), and a time limit.</Paragraph> <Paragraph position="3"> In the right panel of the topic level authoring view, the structure imposed on the data by interaction with the InfoMagnets interface is displayed in sequential form. The top section of the interface (figure 2, section A) has a textbox for specifying an xml file to read. The next section (figure 2, section B), &quot;Move / Rename Subtopic&quot; displays the subtopics. The order of the subtopics displayed in this section acts as a guideline for the agent to follow during the conversation.</Paragraph> <Paragraph position="4"> Double-clicking on a subtopic will display a subtopic view on the right panel. This view acts as a reference for the agent's conversation within the subtopic and is explained in the next section. The author can also rearrange the order of subtopics by selecting a subtopic and using the &quot;>&quot; and &quot;<&quot; buttons to move the subtopic right or left respectively. &quot;x&quot; is used to delete the subtopic. The author can also specify whether the discussion of a subtopic is required (displayed in red) or optional (in green) using the checkbox that is labeled &quot;required&quot;. Clicking on the &quot;Hide Opt&quot; button will only display the required subtopics.</Paragraph> <Paragraph position="5"> The last section of the right panel in topic level authoring view (figure 2, section C) is titled &quot;move subtopic divider&quot;. A blue line denotes the border of the subtopic. The author can move the line up or down to move the boundary of the subtopics automatically inserted by the InfoMagnets interface. The author can also click on any part of conversation and press the &quot;split&quot; button to split the subtopic in two sections. In addition, she can change the label of the sub-topic segment using the drop down list.</Paragraph> <Paragraph position="6"> While the Topic View portion of the authoring interface proper allows specification of which subtopics can occur as part of a dialogue, which are required and which are optional, and what the default ordering is, the Subtopic Level is for specification of the low level turn-by-turn details of what happens within a subtopic segment.</Paragraph> <Paragraph position="7"> This section reports early work on the design of this portion of the interface.</Paragraph> <Paragraph position="8"> The subtopic view displays a structure that the conversational agent refers to in deciding what its next contribution should be. The building blocks from which knowledge sources for the dialogue engine will be generated are templates abstracted from example dialogue segments, similar to KCD specifications (Jordan, Rose, & VanLehn, 2001; Rose & Torry, 2004). As part of the process of abstracting templates, each utterance is tagged with its utterance type using a menu-based interface as in (Gweon et al., submitted). The utterance type determines what would be an appropriate form for a response.</Paragraph> <Paragraph position="9"> Identifying this is meant to allow the dialogue manager to maintain coherence in the emerging dialogue. Users may also trim out undesired portions of text from the actual example fragments in abstracting out templates to be used for generating knowledge sources.</Paragraph> <Paragraph position="10"> Each utterance type has sets of template response types associated with them. The full set of utterance types includes Open questions, Closed questions, Understanding check questions, Assertions, Commands/Requests, Acknowledgements, Acceptances, and Rejections.</Paragraph> <Paragraph position="11"> The templates will not be used in their authored form. Instead, they will be used to generate knowledge sources in the form required by the backend dialogue system as in (Rose & Hall, 2004), although this is still work in progress.</Paragraph> <Paragraph position="12"> Each template is composed of one or more exchanges during which the speaker who initiated the segment maintains conversational control. If control shifts to the other speakers, a new template is used to guide the conversation. After each of the controlling speaker's turns within the segment are listed a number of prototypical responses. One of these responses is a default response that signals that the dialogue should proceed to the next turn in the template. The other prototypical responses are associated with subgoals that are in turn associated with other templates. Thus, the dialogue takes on a hierarchical structure.</Paragraph> <Paragraph position="13"> Mixed initiative interaction is meant to emerge from the underlying template-based structure by means of the multi-threaded discourse management approach discussed in (Rose & Torrey, 2004). To this end, templates are meant to be used in two ways. The first way is when the dialogue system has conversational control. In this case, conversations can be managed as in (Rose et al., 2001). The second way in which templates are used is for determining how to respond when user's have conversational control. Provided that the user's utterances match what is expected of the conversational participant who is in control based on the current template, then the system can simply pick one of the expected responses. Otherwise if at some point the user's response does not match, the system should check whether the user is initiating yet a different segment. If not, then the system should take conversational control.</Paragraph> </Section> class="xml-element"></Paper>