Dialogue Control by Pomdp Using Dialogue Data Statistics

  • Yasuhiro Minami
  • Akira Mori
  • Toyomi Meguro
  • Ryuichiro Higashinaka
  • Kohji Dohsaka
  • Eisaku Maeda
Conference paper

Abstract

Partially Observable Markov Decision Processes (POMDPs) are applied in ac- tion control to manage and support users’ natural dialogue communication with conversational agents. Any agent’s action must be determined, based on probabilistic methods, from noisy data through sensors in the real world. Agents must flexibly choose their actions to reach a target dialogue sequence with the users while retaining as many statistical characteristics of the data as possible. This issue can be solved by two approaches: automatically acquiring POMDP probabilities using Dynamic Bayesian Networks (DBNs)(DBNs) trained from a large amount of dialogue data and obtaining POMDP rewards from human evaluations and agent action predictive probabilities. Using the probabilities and the rewards, POMDP value iteration calculates a policy that can generate an action sequence that maximizes both the predictive distributions of actions and user evaluations.

Keywords

Partially Observable Markov Decision Process (POMDP) Dialogue management DMulti-modal interaction Dynamic Bayesian Network (DBN) Expectation-Maximization (EM) algorithm 

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Copyright information

© Springer Science+Business Media, LLC 2011

Authors and Affiliations

  • Yasuhiro Minami
    • 1
  • Akira Mori
    • 1
  • Toyomi Meguro
    • 1
  • Ryuichiro Higashinaka
    • 2
  • Kohji Dohsaka
    • 1
  • Eisaku Maeda
    • 1
  1. 1.NTT Communication Science Laboratories, NTT CorporationKyotoJapan
  2. 2.NTT Cyber Space Laboratories, NTT CorporationYokosukaJapan

Personalised recommendations