An affect-rich neural conversational model with biased attention and weighted cross-entropy loss
Affect conveys important implicit information in human communication. Having the capability to correctly express affect during human-machine conversations is one of the major milestones in artificial intelligence. In recent years, extensive research on open-domain neural conversational models has be...
Saved in:
Main Authors: | , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2020
|
Subjects: | |
Online Access: | https://arxiv.org/abs/1811.07078 https://hdl.handle.net/10356/139252 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-139252 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1392522020-05-18T06:55:37Z An affect-rich neural conversational model with biased attention and weighted cross-entropy loss Zhong, Peixiang Wang, Di Miao, Chunyan School of Computer Science and Engineering The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19) Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly Alibaba-NTU Singapore Joint Research Institute Engineering::Computer science and engineering Computation and Language Artificial Intelligence Affect conveys important implicit information in human communication. Having the capability to correctly express affect during human-machine conversations is one of the major milestones in artificial intelligence. In recent years, extensive research on open-domain neural conversational models has been conducted. However, embedding affect into such models is still under explored. In this paper, we propose an endto-end affect-rich open-domain neural conversational model that produces responses not only appropriate in syntax and semantics, but also with rich affect. Our model extends the Seq2Seq model and adopts VAD (Valence, Arousal and Dominance) affective notations to embed each word with affects. In addition, our model considers the effect of negators and intensifiers via a novel affective attention mechanism, which biases attention towards affect-rich words in input sentences. Lastly, we train our model with an affect-incorporated objective function to encourage the generation of affect-rich words in the output responses. Evaluations based on both perplexity and human evaluations show that our model outperforms the state-of-the-art baseline model of comparable size in producing natural and affect-rich responses. NRF (Natl Research Foundation, S’pore) MOH (Min. of Health, S’pore) Accepted version 2020-05-18T06:55:37Z 2020-05-18T06:55:37Z 2018 Conference Paper Zhong, P., Wang, D., & Miao, C. (2018). An affect-rich neural conversational model with biased attention and weighted cross-entropy loss. Proceedings of The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19), 7492-7500. https://arxiv.org/abs/1811.07078 https://hdl.handle.net/10356/139252 7492 7500 en © 2019 Association for the Advancement of Artificial Intelligence. All rights reserved. This paper was published in The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19) and is made available with permission of Association for the Advancement of Artificial Intelligence. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
country |
Singapore |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Computation and Language Artificial Intelligence |
spellingShingle |
Engineering::Computer science and engineering Computation and Language Artificial Intelligence Zhong, Peixiang Wang, Di Miao, Chunyan An affect-rich neural conversational model with biased attention and weighted cross-entropy loss |
description |
Affect conveys important implicit information in human communication. Having the capability to correctly express affect during human-machine conversations is one of the major milestones in artificial intelligence. In recent years, extensive research on open-domain neural conversational models has been conducted. However, embedding affect into such models is still under explored. In this paper, we propose an endto-end affect-rich open-domain neural conversational model that produces responses not only appropriate in syntax and semantics, but also with rich affect. Our model extends the Seq2Seq model and adopts VAD (Valence, Arousal and Dominance) affective notations to embed each word with affects. In addition, our model considers the effect of negators and intensifiers via a novel affective attention mechanism, which biases attention towards affect-rich words in input sentences. Lastly, we train our model with an affect-incorporated objective function to encourage the generation of affect-rich words in the output responses. Evaluations based on both perplexity and human evaluations show that our model outperforms the state-of-the-art baseline model of comparable size in producing natural and affect-rich responses. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Zhong, Peixiang Wang, Di Miao, Chunyan |
format |
Conference or Workshop Item |
author |
Zhong, Peixiang Wang, Di Miao, Chunyan |
author_sort |
Zhong, Peixiang |
title |
An affect-rich neural conversational model with biased attention and weighted cross-entropy loss |
title_short |
An affect-rich neural conversational model with biased attention and weighted cross-entropy loss |
title_full |
An affect-rich neural conversational model with biased attention and weighted cross-entropy loss |
title_fullStr |
An affect-rich neural conversational model with biased attention and weighted cross-entropy loss |
title_full_unstemmed |
An affect-rich neural conversational model with biased attention and weighted cross-entropy loss |
title_sort |
affect-rich neural conversational model with biased attention and weighted cross-entropy loss |
publishDate |
2020 |
url |
https://arxiv.org/abs/1811.07078 https://hdl.handle.net/10356/139252 |
_version_ |
1681058366018813952 |