Cargando…
Towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning
PURPOSE: Developing a Dialogue/Virtual Agent (VA) that can handle complex tasks (need) of the user pertaining to multiple intents of a domain is challenging as it requires the agent to simultaneously deal with multiple subtasks. However, majority of these end-to-end dialogue systems incorporate only...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7332012/ https://www.ncbi.nlm.nih.gov/pubmed/32614929 http://dx.doi.org/10.1371/journal.pone.0235367 |
_version_ | 1783553441667219456 |
---|---|
author | Saha, Tulika Saha, Sriparna Bhattacharyya, Pushpak |
author_facet | Saha, Tulika Saha, Sriparna Bhattacharyya, Pushpak |
author_sort | Saha, Tulika |
collection | PubMed |
description | PURPOSE: Developing a Dialogue/Virtual Agent (VA) that can handle complex tasks (need) of the user pertaining to multiple intents of a domain is challenging as it requires the agent to simultaneously deal with multiple subtasks. However, majority of these end-to-end dialogue systems incorporate only user semantics as inputs in the learning process and ignore other useful user behavior and information. Sentiment of the user at the time of conversation plays an important role in securing maximum user gratification. So, incorporating sentiment of the user during the policy learning becomes even more crucial, more so when serving composite tasks of the user. METHODOLOGY: As a first step towards enabling the development of sentiment aided VA for multi-intent conversations, this paper proposes a new dataset, annotated with its corresponding intents, slot and sentiment (considering the entire dialogue history) labels, named SentiVA, collected from open-sourced dialogue datasets. In order to integrate these multiple aspects, a Hierarchical Reinforcement Learning (HRL) specifically options based VA is proposed to learn strategies for managing multi-intent conversations. Along with task success based immediate rewards, sentiment based immediate rewards are also incorporated in the hierarchical value functions to make the VA user adaptive. FINDINGS: Empirically, the paper shows that task based and sentiment based immediate rewards cumulatively are required to ensure successful task completion and attain maximum user satisfaction in a multi-intent scenario instead of any of these rewards alone. PRACTICAL IMPLICATIONS: The eventual evaluators and consumers of dialogue systems are users. Thus, to ensure a fulfilling conversational experience involving maximum user satisfaction requires VA to consider user sentiment at every time-step in its decision making policy. ORIGINALITY: This work is the first attempt in incorporating sentiment based rewards in the HRL framework. |
format | Online Article Text |
id | pubmed-7332012 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-73320122020-07-14 Towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning Saha, Tulika Saha, Sriparna Bhattacharyya, Pushpak PLoS One Research Article PURPOSE: Developing a Dialogue/Virtual Agent (VA) that can handle complex tasks (need) of the user pertaining to multiple intents of a domain is challenging as it requires the agent to simultaneously deal with multiple subtasks. However, majority of these end-to-end dialogue systems incorporate only user semantics as inputs in the learning process and ignore other useful user behavior and information. Sentiment of the user at the time of conversation plays an important role in securing maximum user gratification. So, incorporating sentiment of the user during the policy learning becomes even more crucial, more so when serving composite tasks of the user. METHODOLOGY: As a first step towards enabling the development of sentiment aided VA for multi-intent conversations, this paper proposes a new dataset, annotated with its corresponding intents, slot and sentiment (considering the entire dialogue history) labels, named SentiVA, collected from open-sourced dialogue datasets. In order to integrate these multiple aspects, a Hierarchical Reinforcement Learning (HRL) specifically options based VA is proposed to learn strategies for managing multi-intent conversations. Along with task success based immediate rewards, sentiment based immediate rewards are also incorporated in the hierarchical value functions to make the VA user adaptive. FINDINGS: Empirically, the paper shows that task based and sentiment based immediate rewards cumulatively are required to ensure successful task completion and attain maximum user satisfaction in a multi-intent scenario instead of any of these rewards alone. PRACTICAL IMPLICATIONS: The eventual evaluators and consumers of dialogue systems are users. Thus, to ensure a fulfilling conversational experience involving maximum user satisfaction requires VA to consider user sentiment at every time-step in its decision making policy. ORIGINALITY: This work is the first attempt in incorporating sentiment based rewards in the HRL framework. Public Library of Science 2020-07-02 /pmc/articles/PMC7332012/ /pubmed/32614929 http://dx.doi.org/10.1371/journal.pone.0235367 Text en © 2020 Saha et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Saha, Tulika Saha, Sriparna Bhattacharyya, Pushpak Towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning |
title | Towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning |
title_full | Towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning |
title_fullStr | Towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning |
title_full_unstemmed | Towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning |
title_short | Towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning |
title_sort | towards sentiment aided dialogue policy learning for multi-intent conversations using hierarchical reinforcement learning |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7332012/ https://www.ncbi.nlm.nih.gov/pubmed/32614929 http://dx.doi.org/10.1371/journal.pone.0235367 |
work_keys_str_mv | AT sahatulika towardssentimentaideddialoguepolicylearningformultiintentconversationsusinghierarchicalreinforcementlearning AT sahasriparna towardssentimentaideddialoguepolicylearningformultiintentconversationsusinghierarchicalreinforcementlearning AT bhattacharyyapushpak towardssentimentaideddialoguepolicylearningformultiintentconversationsusinghierarchicalreinforcementlearning |