论文标题
与异质图神经网络的情感对话产生
Emotional Conversation Generation with Heterogeneous Graph Neural Network
论文作者
论文摘要
成功的情感对话系统取决于足够的感知和对情绪的适当表达。在现实生活中,人类首先本能地从多源信息中观察到情感,包括对话历史上隐藏的情感流,面部表情,音频和演讲者的个性。然后,他们根据自己的个性传达适当的情绪,但是这些多种类型的信息在情感对话领域不足。为了解决这个问题,在本文中,我们提出了一种基于情感对话的基于图形的模型。首先,我们设计了一个基于图的编码器,以用异质的图神经网络表示对话内容(即对话历史,其情感流,面部表情,音频和说话者的个性),然后预测适当的情感以提供反馈。其次,我们采用情感人物感知的解码器来产生与对话背景以及适当情绪相关的响应,通过采用编码的图表表示,编码器的预测情绪以及当前说话者的个性作为输入。自动评估和人类评估的实验表明,我们的方法可以有效地感知来自多源知识的情绪,并产生令人满意的响应。此外,基于最新的文本生成器Bart,我们的模型仍然可以实现一致的改进,这极大地超过了一些现有的最新模型。
The successful emotional conversation system depends on sufficient perception and appropriate expression of emotions. In a real-life conversation, humans firstly instinctively perceive emotions from multi-source information, including the emotion flow hidden in dialogue history, facial expressions, audio, and personalities of speakers. Then, they convey suitable emotions according to their personalities, but these multiple types of information are insufficiently exploited in emotional conversation fields. To address this issue, in this paper, we propose a heterogeneous graph-based model for emotional conversation generation. Firstly, we design a Heterogeneous Graph-Based Encoder to represent the conversation content (i.e., the dialogue history, its emotion flow, facial expressions, audio, and speakers' personalities) with a heterogeneous graph neural network, and then predict suitable emotions for feedback. Secondly, we employ an Emotion-Personality-Aware Decoder to generate a response relevant to the conversation context as well as with appropriate emotions, through taking the encoded graph representations, the predicted emotions by the encoder and the personality of the current speaker as inputs. Experiments on both automatic and human evaluation show that our method can effectively perceive emotions from multi-source knowledge and generate a satisfactory response. Furthermore, based on the up-to-date text generator BART, our model still can achieve consistent improvement, which significantly outperforms some existing state-of-the-art models.