Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
An Overview of MAXQ Hierarchical Reinforcement Learning
SARA '02 Proceedings of the 4th International Symposium on Abstraction, Reformulation, and Approximation
Modelling Navigational Knowledge by Route Graphs
Spatial Cognition II, Integrating Abstract Theories, Empirical Studies, Formal Methods, and Practical Applications
Elements of Good Route Directions in Familiar and Unfamiliar Environments
COSIT '99 Proceedings of the International Conference on Spatial Information Theory: Cognitive and Computational Foundations of Geographic Information Science
The Nature of Landmarks for Real and Electronic Spaces
COSIT '99 Proceedings of the International Conference on Spatial Information Theory: Cognitive and Computational Foundations of Geographic Information Science
Enriching Wayfinding Instructions with Local Landmarks
GIScience '02 Proceedings of the Second International Conference on Geographic Information Science
Artificial Intelligence: A Modern Approach
Artificial Intelligence: A Modern Approach
Towards developing general models of usability with PARADISE
Natural Language Engineering
Enabling technology for multilingual natural language generation: the KPML development environment
Natural Language Engineering
Adaptive navigation support with public displays
Proceedings of the 10th international conference on Intelligent user interfaces
Building deep dependency structures with a wide-coverage CCG parser
ACL '02 Proceedings of the 40th Annual Meeting on Association for Computational Linguistics
Tractable planning under uncertainty: exploiting structure
Tractable planning under uncertainty: exploiting structure
Probabilistic policy reuse in a reinforcement learning agent
AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
The Knowledge Engineering Review
Simplest Instructions: Finding Easy-to-Describe Routes for Navigation
GIScience '08 Proceedings of the 5th international conference on Geographic Information Science
Hybrid reinforcement/supervised learning of dialogue policies from fixed data sets
Computational Linguistics
Natural Language Engineering
Daisie: Information State Dialogues for Situated Systems
TSD '09 Proceedings of the 12th International Conference on Text, Speech and Dialogue
The Hidden Information State model: A practical framework for POMDP-based spoken dialogue management
Computer Speech and Language
Evaluation of a hierarchical reinforcement learning spoken dialogue system
Computer Speech and Language
Journal of Artificial Intelligence Research
Hierarchical reinforcement learning with the MAXQ value function decomposition
Journal of Artificial Intelligence Research
Optimizing dialogue management with reinforcement learning: experiments with the NJFun system
Journal of Artificial Intelligence Research
Reinforcement learning: a survey
Journal of Artificial Intelligence Research
Learning to adapt to unknown users: referring expression generation in spoken dialogue systems
ACL '10 Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics
Hierarchical reinforcement learning for adaptive text generation
INLG '10 Proceedings of the 6th International Natural Language Generation Conference
Generating adaptive route instructions using hierarchical reinforcement learning
SC'10 Proceedings of the 7th international conference on Spatial cognition
Fast reinforcement learning of dialogue policies using stable function approximation
IJCNLP'04 Proceedings of the First international joint conference on Natural Language Processing
Enabling Multimodal Human–Robot Interaction for the Karlsruhe Humanoid Robot
IEEE Transactions on Robotics
ENLG '11 Proceedings of the 13th European Workshop on Natural Language Generation
The Bremen system for the GIVE-2.5 challenge
ENLG '11 Proceedings of the 13th European Workshop on Natural Language Generation
Comparing HMMs and Bayesian networks for surface realisation
NAACL HLT '12 Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
Optimising incremental dialogue decisions using information density for interactive systems
EMNLP-CoNLL '12 Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning
Machine learning for interactive systems and robots: a brief introduction
Proceedings of the 2nd Workshop on Machine Learning for Interactive Systems: Bridging the Gap Between Perception, Action and Communication
Hi-index | 0.00 |
This article addresses the problem of scalable optimization for spatially-aware dialogue systems. These kinds of systems must perceive, reason, and act about the spatial environment where they are embedded. We formulate the problem in terms of Semi-Markov Decision Processes and propose a hierarchical reinforcement learning approach to optimize subbehaviors rather than full behaviors. Because of the vast number of policies that are required to control the interaction in a dynamic environment (e.g., a dialogue system assisting a user to navigate in a building from one location to another), our learning approach is based on two stages: (a) the first stage learns low-level behavior, in advance; and (b) the second stage learns high-level behavior, in real time. For such a purpose we extend an existing algorithm in the literature of reinforcement learning in order to support reusable policies and therefore to perform fast learning. We argue that our learning approach makes the problem feasible, and we report on a novel reinforcement learning dialogue system that performs a joint optimization between dialogue and spatial behaviors. Our experiments, using simulated and real environments, are based on a text-based dialogue system for indoor navigation. Experimental results in a realistic environment reported an overall user satisfaction result of 89%, which suggests that our proposed approach is attractive for its application in real interactions as it combines fast learning with adaptive and reasonable behavior.