An affect-rich neural conversational model with biased attention and weighted cross-entropy loss

Affect conveys important implicit information in human communication. Having the capability to correctly express affect during human-machine conversations is one of the major milestones in artificial intelligence. In recent years, extensive research on open-domain neural conversational models has be...

Full description

Saved in:
Bibliographic Details
Main Authors: Zhong, Peixiang, Wang, Di, Miao, Chunyan
Other Authors: School of Computer Science and Engineering
Format: Conference or Workshop Item
Language:English
Published: 2020
Subjects:
Online Access:https://arxiv.org/abs/1811.07078
https://hdl.handle.net/10356/139252
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-139252
record_format dspace
spelling sg-ntu-dr.10356-1392522020-05-18T06:55:37Z An affect-rich neural conversational model with biased attention and weighted cross-entropy loss Zhong, Peixiang Wang, Di Miao, Chunyan School of Computer Science and Engineering The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19) Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly Alibaba-NTU Singapore Joint Research Institute Engineering::Computer science and engineering Computation and Language Artificial Intelligence Affect conveys important implicit information in human communication. Having the capability to correctly express affect during human-machine conversations is one of the major milestones in artificial intelligence. In recent years, extensive research on open-domain neural conversational models has been conducted. However, embedding affect into such models is still under explored. In this paper, we propose an endto-end affect-rich open-domain neural conversational model that produces responses not only appropriate in syntax and semantics, but also with rich affect. Our model extends the Seq2Seq model and adopts VAD (Valence, Arousal and Dominance) affective notations to embed each word with affects. In addition, our model considers the effect of negators and intensifiers via a novel affective attention mechanism, which biases attention towards affect-rich words in input sentences. Lastly, we train our model with an affect-incorporated objective function to encourage the generation of affect-rich words in the output responses. Evaluations based on both perplexity and human evaluations show that our model outperforms the state-of-the-art baseline model of comparable size in producing natural and affect-rich responses. NRF (Natl Research Foundation, S’pore) MOH (Min. of Health, S’pore) Accepted version 2020-05-18T06:55:37Z 2020-05-18T06:55:37Z 2018 Conference Paper Zhong, P., Wang, D., & Miao, C. (2018). An affect-rich neural conversational model with biased attention and weighted cross-entropy loss. Proceedings of The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19), 7492-7500. https://arxiv.org/abs/1811.07078 https://hdl.handle.net/10356/139252 7492 7500 en © 2019 Association for the Advancement of Artificial Intelligence. All rights reserved. This paper was published in The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19) and is made available with permission of Association for the Advancement of Artificial Intelligence. application/pdf
institution Nanyang Technological University
building NTU Library
country Singapore
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Computation and Language
Artificial Intelligence
spellingShingle Engineering::Computer science and engineering
Computation and Language
Artificial Intelligence
Zhong, Peixiang
Wang, Di
Miao, Chunyan
An affect-rich neural conversational model with biased attention and weighted cross-entropy loss
description Affect conveys important implicit information in human communication. Having the capability to correctly express affect during human-machine conversations is one of the major milestones in artificial intelligence. In recent years, extensive research on open-domain neural conversational models has been conducted. However, embedding affect into such models is still under explored. In this paper, we propose an endto-end affect-rich open-domain neural conversational model that produces responses not only appropriate in syntax and semantics, but also with rich affect. Our model extends the Seq2Seq model and adopts VAD (Valence, Arousal and Dominance) affective notations to embed each word with affects. In addition, our model considers the effect of negators and intensifiers via a novel affective attention mechanism, which biases attention towards affect-rich words in input sentences. Lastly, we train our model with an affect-incorporated objective function to encourage the generation of affect-rich words in the output responses. Evaluations based on both perplexity and human evaluations show that our model outperforms the state-of-the-art baseline model of comparable size in producing natural and affect-rich responses.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Zhong, Peixiang
Wang, Di
Miao, Chunyan
format Conference or Workshop Item
author Zhong, Peixiang
Wang, Di
Miao, Chunyan
author_sort Zhong, Peixiang
title An affect-rich neural conversational model with biased attention and weighted cross-entropy loss
title_short An affect-rich neural conversational model with biased attention and weighted cross-entropy loss
title_full An affect-rich neural conversational model with biased attention and weighted cross-entropy loss
title_fullStr An affect-rich neural conversational model with biased attention and weighted cross-entropy loss
title_full_unstemmed An affect-rich neural conversational model with biased attention and weighted cross-entropy loss
title_sort affect-rich neural conversational model with biased attention and weighted cross-entropy loss
publishDate 2020
url https://arxiv.org/abs/1811.07078
https://hdl.handle.net/10356/139252
_version_ 1681058366018813952