A Two-tier User Simulation Model for Reinforcement Learning of Adaptive Referring Expression Generation Policies Srinivasan Janarthanam School of Informatics University of Edinburgh [email protected]

Abstract We present a new two-tier user simulation model for learning adaptive referring expression generation (REG) policies for spoken dialogue systems using reinforcement learning. Current user simulation models that are used for dialogue policy learning do not simulate users with different levels of domain expertise and are not responsive to referring expressions used by the system. The twotier model displays these features, that are crucial to learning an adaptive REG policy. We also show that the two-tier model simulates real user behaviour more closely than other baseline models, using the dialogue similarity measure based on KullbackLeibler divergence.

1

Introduction

We present a new user simulation model for learning adaptive referring expression generation (REG) policies for spoken dialogue systems using reinforcement learning methods. An adaptive REG policy equips a dialogue system to dynamically modify its utterances in order to adapt to user’s domain knowledge level. For instance, to refer to the domain objects, the system might use simple descriptive expressions with novices and technical jargon with experts. Such adaptations help grounding between the dialogue partners (Issacs and Clark, 1987). Since the user’s knowledge level is unknown, the system must be able to adapt dynamically during the conversation. Hand-coding such a policy could be extremely difficult. (Janarthanam and Lemon, 2009b)

Oliver Lemon School of Informatics University of Edinburgh [email protected]

have shown that such policies can be learned using simulation based reinforcement learning (RL) methods. The quality of such learned policies is directly dependent on the performance of the user simulations used to train them. So far, only hand-coded user simulations have been employed. In contrast, we now present a data driven two-tier user simulation model trained on dialogue data collected from real users. We also show that the two-tier model simulates real users more faithfully than other data driven baseline n-gram models (Eckert et al., 1997). In section 2 , we briefly discuss other work related to user simulations for dialogue policy learning using RL. In section 3, we describe the data used to build the simulation. Section 4 describes in detail the simulation models. In section 5 and 6, we present the evaluation metrics used and the results.

2 Related work Several user simulation models have been proposed for dialogue management policy learning (Schatzmann et al., 2006; Schatzmann et al., 2007). However, these models cannot be directly used for REG policy learning because they interact with the dialogue system only using high-level dialogue acts. In order to learn REG policies, user simulations need to respond to the system’s choice of referring expressions. Real users request clarification when they do not understand the referring expressions used by a system. Also, to learn an adaptive policy, the user simulations need to simulate user groups with different knowledge levels, a feature which the current models do not support. We propose a two-tier simulation which simulates users with different knowl-

edge levels and is sensitive to the system’s choice of referring expressions.

pressions RECs,t and output the user’s dialogue act Au,t and environment act EAu,t .

3

4.1

Corpus

The “Wizard-of-Oz” (WOZ) methodology is a widely accepted way of collecting dialogue data for user simulation modeling (Whittaker et al., 2002). Usually, in a Wizard-of-Oz setup, real users interact with a human wizard disguised as a dialogue system. The wizard decides optimal responses based on the current dialogue state. This interaction is recorded and the data is used to train user simulations using supervised learning methods. Using a WOZ setup, we collected dialogue data from 17 participants. However, in our setup, the wizard plays the role of only an interpreter and does not make any decisions, and we collect data on user reactions to system REG decisions. There is a dialogue system that decides the responses to user’s moves. The dialogue task of the participant is to interact with the dialogue system to get instructions to setup a broadband Internet connection. The referring expression generation strategy is chosen before the dialogue starts and stays the same for the whole session. The strategies used were “jargon”, “descriptive” and “tutorial”. In the jargon strategy the system instructs the user using technical terms (e.g. “Plug the broadband filter into the phone socket.”). In the descriptive strategy, it uses descriptive terms (e.g. “Plug the small white box into the square white box on the wall.”). In the tutorial strategy, the system uses both jargon and descriptive terms together. The system provides clarifications on referring expressions when users request them. Please refer to (Janarthanam and Lemon, 2009a) for a more details on our Wizard-of-Oz environment for data collection. The participant’s ability to recognise domain objects using technical terms were recorded before and after the dialogue task. The dialogues were around 24 to 35 turns long depending on the strategy and user’s domain knowledge.

4

User Simulation models

The dialogue data and knowledge profiles collected from the participants were used to build user simulation models. These models take as input the system’s dialogue act As,t and choice of referring ex-

Advanced n-gram model

A simple approach to model real user behaviour is to model user responses (dialogue act and environment act) on many context variables - all referring expressions used in the utterance, the user’s knowledge of the REs, history of clarification requests on the REs, and the system’s dialogue act. With such a large context, the advanced n-gram model (Georgila et al., 2006) is ideally the probability distribution found in the real user data. P (Au,t |As,t , RECs,t , DKu , H) P (EAu,t |As,t , RECs,t , DKu , H) However, with such complex contexts, there are data sparsity problems because very many contexts are not seen in the small amount of collected WOZ data. 4.2

A Two-tier model

Instead of such a complex context model, we could backoff and use smaller subcontexts derived from it. We propose a two-tier model, in which the simulation of a user’s response is divided into two steps. In the first step, the simulation processes all the referring expressions used by the system RECs,t . Unlike the advanced n-gram model, this is done one by one, for each expression REs,t separately and not for the whole set RECs,t at once. It returns a clarification request based on each referring expression REs,t used, the user’s knowledge of the expression DKRE,u , and previous clarification requests on the expression HRE and the system dialogue act As,t . The clarification request is highly likely in case of the jargon strategy and less likely in other strategies. Also, if a clarification has already been issued, the user is less likely to issue another request for clarification. In such cases, the clarification request model returns none. P (CRu,t |REs,t , DKRE,u , HRE , As,t ) In the next step, the model returns a user dialogue act Au,t and an environment act EAu,t based on the system dialogue act As,t and the clarification request CRu,t . P (Au,t |As,t , CRu,t )

P ∗ (ei ) =

P (EAu,t |As,t , CRu,t ) By dividing the rich context into smaller subcontexts between the two steps, the two-tier model simulates real users in contexts that are not directly observed in the dialogue data. The model will therefore respond to system utterances containing a mix of REG strategies (for e.g. one jargon and one descriptive expression in the same utterance). 4.3 Baseline Bigram model A bigram model was built using the dialogue data by conditioning the user responses only on the system’s dialogue act (Eckert et al., 1997). P (Au,t |As,t ) P (EAu,t |As,t ) Since it ignores all the context variables except the system dialogue act, it can be used in contexts that are not observed in the dialogue data. 4.4 Trigram model The trigram model is similar to the bigram model, but with the previous system dialogue act As,t−1 as an additional context variable. P (Au,t |As,t , As,t−1 ) P (EAu,t |As,t , As,t−1 ) 4.5 Equal Probability model baseline The equal probability model is similar to the bigram model, except that it is not trained on the dialogue data. Instead, it assigns equal probability to all possible responses for the given system dialogue act. 4.6 Smoothing We used Witten-Bell discounting to smooth all our models except the equal probability model, in order to account for unobserved but possible events in dialogue contexts. Witten-Bell discounting extracts a small percentage of probability mass, i.e. number of distinct events observed for the first time (T ) in a context, out of the total number of instances (N ), and redistributes this mass to unobserved events in the given context (V − T ) (where V is the number of all possible events) . The discounted probabilities P ∗ of observed events (C(ei ) > 0) and unobserved events (C(ei ) = 0) are given below. P ∗ (ei ) =

C(ei ) N +T

if (C(ei ) > 0)

t (N +T )(V −T )

if (C(ei ) = 0)

On analysis, we found that the Witten-Bell discounting assigns greater probability to unobserved events than to observed events, in cases where the number of events per context is very low. For instance, in a particular context, the possible events, their frequencies and their original probabilities were - provide info (3, 0.75), other (1, 0.25), request clarification (0, 0). After discounting, the revised probabilities P ∗ are 0.5, 0.167 and 0.33. request clarification gets the whole share of extracted probability as it is the only unobserved event in the context and is more than the other events actually observed in the data. This is counter-intuitive for our application. Therefore, we use a modified version of Witten-Bell discounting (given below) to smooth our models, where the extracted probability is equally divided amongst all possible events. Using the modified version, the revised probabilities for the illustrated example are 0.61, 0.28 and 0.11 respectively. P ∗ (ei ) =

C(ei ) N +T

+

T (N +T )V

5 Metrics for evaluation of simulations While there are many proposed measures to rank user simulation models with respect to real user data (Schatzmann et al., 2005; Georgila et al., 2006; Rieser and Lemon, 2006a; Williams, 2008), we use the Dialogue Similarity measure based on Kullback-Leibler (KL) (Cuayahuitl et al., 2005; Cuayahuitl, 2009) divergence to measure how similar the probability distributions of the simulation models are to the real human data. 5.1

Dialogue Similarity

Dialogue Similarity is a measure of divergence between real and simulated dialogues and can measure how similar a model is to real data. The measure is based on Kullback-Leibler (KL) divergence and is defined as follows: DS(P ||Q) =

1 N

PN

DKL (P ||Q)+DKL (Q||P ) 2 PM pi i=1 pi ∗ log( qi )

i=1

DKL (P ||Q) =

The metric measures the divergence between distributions P and Q in N different contexts with M responses per context. Ideally, the dialogue similarity between two similar distributions is close to zero.

Model Two-tier Bigram Trigram Equal Probability

Au,t 0.078 0.150 0.145 0.445

EAu,t 0.018 0.139 0.158 0.047

Table 1: Dialogue Similarity with Modified WittenBell discounting

6

Evaluation results

We consider the advanced n-gram model to be a realistic model of the corpus, as it is takes into account all context variables and is reasonably smoothed to account for unobserved responses. Therefore, we compare the probability distributions of all the other models to the advanced n-gram model using the dialogue similarity measure. The results of the evaluation are given in table 1. The results show that the two-tier model is much closer (0.078, 0.018) to the real user data than the other models. This is due to the fact that the bigram and trigram models don’t take into account the factors like the user’s knowledge, the strategy used, and the dialogue history. By effectively dividing the RE processing and the environment interaction, the two-tier simulation model is not only realistic in observed contexts but also usable in unobserved contexts.

7

Conclusion

We have presented a data driven user simulation model called the two-tier model for learning REG policies using reinforcement learning. We have also shown that the two-tier model is much closer to real user data than the other baseline models. We will now train REG policies using the two-tier model and test them on real users in the future.

Acknowledgements The research leading to these results has received funding from the EPSRC (project no. EP/E019501/1) and from the European Community’s Seventh Framework Programme (FP7/20072013) under grant agreement no. 216594 (CLASSiC project www.classic-project.org), and from the British Council’s UKERI programme.

References H. Cuayahuitl, S. Renals, O. Lemon, and H. Shimodaira. 2005. Human-Computer Dialogue Simulation Using Hidden Markov Models. In Proc. of ASRU 2005. H. Cuayahuitl. 2009. Hierarchical Reinforcement Learning for Spoken Dialogue Systems. Ph.D. thesis, University of Edinburgh, UK. W. Eckert, E. Levin, and R. Pieraccini. 1997. User Modeling for Spoken Dialogue System Evaluation. In Proc. of ASRU97. K. Georgila, J. Henderson, and O. Lemon. 2006. User Simulation for Spoken Dialogue System: Learning and Evaluation. In Proc of ICSLP 2006. E. A. Issacs and H. H. Clark. 1987. References in conversations between experts and novices. Journal of Experimental Psychology: General, 116:26–37. S. Janarthanam and O. Lemon. 2009a. A Wizard-ofOz environment to study Referring Expression Generation in a Situated Spoken Dialogue Task. In Proc. ENLG’09. S. Janarthanam and O. Lemon. 2009b. Learning Lexical Alignment Policies for Generating Referring Expressions for Spoken Dialogue Systems. In Proc. ENLG’09. V. Rieser and O. Lemon. 2006a. Cluster-based User Simulations for Learning Dialogue Strategies. In Proc. Interspeech/ICSLP. J. Schatzmann, K. Georgila, and S. J. Young. 2005. Quantitative Evaluation of User Simulation Techniques for Spoken Dialogue Systems. In Proc. SIGdial workshop on Discourse and Dialogue ’05. J. Schatzmann, K. Weilhammer, M. N. Stuttle, and S. J. Young. 2006. A Survey of Statistical User Simulation Techniques for Reinforcement Learning of Dialogue Management Strategies. Knowledge Engineering Review, pages 97–126. J. Schatzmann, B. Thomson, K. Weilhammer, H. Ye, and S. J. Young. 2007. Agenda-based User Simulation for Bootstrapping a POMDP Dialogue System. In Proc of HLT/NAACL 2007. S. Whittaker, M. Walker, and J. Moore. 2002. Fish or Fowl: A Wizard of Oz Evaluation of Dialogue Strategies in the Restaurant Domain. In Language Resources and Evaluation Conference. J. Williams. 2008. Evaluating User Simulations with the Cramer-von Mises Divergence. Speech Communication, 50:829–846.

A Two-tier User Simulation Model for Reinforcement ...

policies for spoken dialogue systems using rein- forcement ... dialogue partners (Issacs and Clark, 1987). ... and we collect data on user reactions to system REG.

132KB Sizes 0 Downloads 277 Views

Recommend Documents

Bipartite Graph Reinforcement Model for Web Image ...
Sep 28, 2007 - retrieving abundant images on the internet. ..... 6.4 A Comparison with HITS Algorithm ... It is a link analysis algorithm that rates web pages.

A Theory of Model Selection in Reinforcement Learning - Deep Blue
seminar course is my favorite ever, for introducing me into statistical learning the- ory and ..... 6.7.2 Connections to online learning and bandit literature . . . . 127 ...... be to obtain computational savings (at the expense of acting suboptimall

Modelling & Simulation (A Model Curriculum for High Schools ...
Whoops! There was a problem loading more pages. Main menu. Displaying Modelling & Simulation (A Model Curriculum for High Schools) - Norfolk, Virginia.pdf.

SIMULATION - OPTIMIZATION MODEL FOR DOKAN RESERVOIR ...
SIMULATION - OPTIMIZATION MODEL FOR DOKAN RESERVOIR SYSTEM OPERATION.pdf. SIMULATION - OPTIMIZATION MODEL FOR DOKAN ...

A Theory of Model Selection in Reinforcement Learning
4.1 Comparison of off-policy evaluation methods on Mountain Car . . . . . 72 ..... The base of log is e in this thesis unless specified otherwise. To verify,. γH Rmax.

A Multi-layer Model for Face Aging Simulation - Springer Link
ment cosmetic surgery planning, age-adaptive human computer interaction, etc. ... support vector machines (SVMs) to predict the shape in the future. ... [9] defined both a person-specific and a global aging axis in the shape and texture ...

A Simulation Based Model Checker for Real Time Java.pdf ...
checkers can also deal with liveness properties, e.g., by check- ing assertions expressed in linear time logic (LTL) [11]. Figure 1: JPF architecture. Java PathFinder is an explicit state model checker for. Java bytecode. JPF focuses on finding bugs

COTC02: a cotton growth simulation model for global ...
growth, development, yield and water use has been constructed. It is named ..... ENVIRN. DARCY. Darcian flow of water among soil cells (surface energy.

Toward a Model of Mobile User Engagement
conceptual framework for studying mobile information interaction. ... aesthetic appeal, and endurability for the mobile environment. .... mobile devices for social networking purposes. However, .... of perceived visual aesthetics of Web sites.

simulation model pdf
Page 1 of 1. File: Simulation model pdf. Download now. Click here if your download doesn't start automatically. Page 1 of 1. simulation model pdf. simulation model pdf. Open. Extract. Open with. Sign In. Main menu. Displaying simulation model pdf. Pa

Parameter estimation for agenda-based user simulation
In spoken dialogue systems research, modelling .... Each time the user simulator receives a system act, .... these systems are commonly called slot-filling di-.

User Simulation for Online Adaptation and Knowledge ...
Troubleshooting Dialogue Systems ... Current dialogue systems interact with homogeneous user .... Simulates a modem, ADSL filters, a computer , phone line,.

User Simulation for Online Adaptation and Knowledge ...
Troubleshooting domain: Dialogue system helps users ... They do not focus on user's domain knowledge, frustration or .... When do users get frustrated?

PAC Reinforcement Learning with an Imperfect Model
than in M for all X ⊆ X ξ-inc . (see the agnostic version of the conditions in the paper.) Theorem 1: Under Condition 1, there exists an algorithm that achieves. O(|X ξ-inc. |. 2. H. 4 log(1/δ)/ε. 3. ) sample complexity for ξ = O(ε/H. 2. ). A

PAC Reinforcement Learning with an Imperfect Model
The pseudocode is given in Algorithm 1. We first walk through the pseudocode and give some intuitions, and then state and prove the sample complexity guarantee. The outer loop of the algorithm computes πt = π⋆. Mt , the optimal policy of the curr

Pressure Model of Soft Body Simulation
problems appear with the Green function approach (see [6]) - com- plexity of implementation is ... model further, we can ”close the object” and put the ”wind source” within it. .... vector are multiplied by scalar coefficient to get effects o

A Website Mining Model Centered on User Queries
Web servers register important data about the usage of a website. This in- formation generally ... Our model also generates a visualization of the site's content ...

User-directed Non-Disruptive Topic Model Update for ...
lication database to classify the collection of papers into three topics: Natural Language Processing (NLP), ... algorithms to address these basic usability issues. Our work is the first in this direction. Topic Model ... some use Variational Bayes (

User Message Model: A New Approach to Scalable ...
z Nz|u; Nw|z the number of times word w has been assigned to topic z, and N·|z = ∑ w Nw|z; (·) zu mn the count that does not include the current assignment of zu mn. Figure 2 gives the pseudo code for a single Gibbs iteration. After obtaining the

A Website Mining Model Centered on User Queries
Internal queries: These are queries submitted to a website's internal search box. Additionally, external queries that are specified by users for a partic- ular site, will be considered as internal queries for that site. For example,. Google.com queri

SNIF-ACT: A Cognitive Model of User Navigation ... - Semantic Scholar
Applications of the SNIF-ACT Model. 6.2. Cognitive Models ..... The model also requires the specification of the attentional weight parame- ter Wj. We have ..... Parse the Interface Objects, Coded Protocol, and Event Log to deter- mine the next ...

SNIF-ACT: A Cognitive Model of User Navigation ... - Semantic Scholar
Users With Different Background Knowledge. APPENDIX A. THE ... SNIF-ACT. 357. 1. Internet use is estimated to be 68.3% of the North American population.

Web Application Model Recovery for User Input ...
In the internet era, Web applications are becoming the core business in many areas. Meanwhile, there is ..... Computer Society, Toronto, Ontario, 2001, pp.25-34.

Reinforcement Learning as a Context for Integrating AI ...
placing it at a low level would provide maximum flexibility in simulations. Furthermore ... long term achievement of values. It is important that powerful artificial ...