User feedback can be an effective indicator to the success of the human-robot conversation. However, to avoid to interrupt the online real-time conversation process, explicit feedback is usually gained at the end of a conversation. Alternatively, users' responses usually contain their implicit feedback, such as stance, sentiment, emotion, etc., towards the conversation content or the interlocutors. Therefore, exploring the implicit feedback is a natural way to optimize the conversation generation process. In this paper, we propose a novel reward function which explores the implicit feedback to optimize the future reward of a reinforcement learning based neural conversation model. A simulation strategy is applied to explore the state-action ...
Abstract — In order to be useful in real-world situations, it is critical to allow non-technical use...
The synchronization of words in conversation, called entrainment, is generally observed in human-h...
Axelsson A, Buschmeier H, Skantze G. Modeling feedback in interaction with conversational agents—A r...
The ability to compute an accurate reward function is essential for optimising a dialogue policy via...
To train a statistical spoken dialogue system (SDS) it is essential that an accurate method for meas...
This paper develops a Chatbot conversational model that is aimed to achieve two goals: 1) utilizing ...
Viewing dialogue management as a reinforcement learning task enables a system to learn to act optima...
Open-domain dialog generation is a challenging problem; maximum likelihood training can lead to repe...
© 2019 Neural information processing systems foundation. All rights reserved. Building an open-domai...
Statistical spoken dialogue systems have the attractive property of being able to be optimised from ...
Reactions such as gestures and facial expressions are an abundant, natural source of signal emitted ...
In spoken dialogue systems, we aim to deploy artificial intelligence to build automated dialogue age...
We explore unconstrained natural language feedback as a learning signal for artificial agents. Human...
Building a controllable neural conversation model (NCM) is an important task. In this paper, we focu...
The performance of adversarial dialogue generation models relies on the quality of the reward signal...
Abstract — In order to be useful in real-world situations, it is critical to allow non-technical use...
The synchronization of words in conversation, called entrainment, is generally observed in human-h...
Axelsson A, Buschmeier H, Skantze G. Modeling feedback in interaction with conversational agents—A r...
The ability to compute an accurate reward function is essential for optimising a dialogue policy via...
To train a statistical spoken dialogue system (SDS) it is essential that an accurate method for meas...
This paper develops a Chatbot conversational model that is aimed to achieve two goals: 1) utilizing ...
Viewing dialogue management as a reinforcement learning task enables a system to learn to act optima...
Open-domain dialog generation is a challenging problem; maximum likelihood training can lead to repe...
© 2019 Neural information processing systems foundation. All rights reserved. Building an open-domai...
Statistical spoken dialogue systems have the attractive property of being able to be optimised from ...
Reactions such as gestures and facial expressions are an abundant, natural source of signal emitted ...
In spoken dialogue systems, we aim to deploy artificial intelligence to build automated dialogue age...
We explore unconstrained natural language feedback as a learning signal for artificial agents. Human...
Building a controllable neural conversation model (NCM) is an important task. In this paper, we focu...
The performance of adversarial dialogue generation models relies on the quality of the reward signal...
Abstract — In order to be useful in real-world situations, it is critical to allow non-technical use...
The synchronization of words in conversation, called entrainment, is generally observed in human-h...
Axelsson A, Buschmeier H, Skantze G. Modeling feedback in interaction with conversational agents—A r...