论文标题
部分可观测时空混沌系统的无模型预测
Explaining (Sarcastic) Utterances to Enhance Affect Understanding in Multimodal Dialogues
论文作者
论文摘要
对话成为交换思想和概念的主要媒体。从听众的角度来看,确定各种情感素质,例如讽刺,幽默和情感,对于理解发出的话语的真实含义至关重要。但是,学习这些影响维度面临的主要障碍之一是具有象征性语言的存在,即。具有讽刺意味的,隐喻或讽刺。我们假设,任何构成发出话语的详尽和明确表示的检测系统都将改善对话的整体理解。为此,我们探讨了对话中讽刺解释的任务,该任务旨在展现讽刺性话语背后的隐藏讽刺。我们提出了一个深层神经网络摩西,该网络将多模式(讽刺)对话实例作为输入,并生成自然语言句子作为其解释。随后,我们利用了生成的解释来在对话对话设置中进行各种自然语言理解任务,例如讽刺检测,幽默识别和情感识别。我们的评估表明,在不同的评估指标(例如鲁日,bleu和流星)上,摩西的表现平均超过了SED的最新系统。此外,我们观察到,利用生成的解释前进了三个下游任务来影响分类 - 在讽刺检测任务中,平均提高了〜14%的F1得分,而在幽默识别和情感识别任务中,平均提高了〜2%。我们还进行了广泛的分析以评估结果的质量。
Conversations emerge as the primary media for exchanging ideas and conceptions. From the listener's perspective, identifying various affective qualities, such as sarcasm, humour, and emotions, is paramount for comprehending the true connotation of the emitted utterance. However, one of the major hurdles faced in learning these affect dimensions is the presence of figurative language, viz. irony, metaphor, or sarcasm. We hypothesize that any detection system constituting the exhaustive and explicit presentation of the emitted utterance would improve the overall comprehension of the dialogue. To this end, we explore the task of Sarcasm Explanation in Dialogues, which aims to unfold the hidden irony behind sarcastic utterances. We propose MOSES, a deep neural network, which takes a multimodal (sarcastic) dialogue instance as an input and generates a natural language sentence as its explanation. Subsequently, we leverage the generated explanation for various natural language understanding tasks in a conversational dialogue setup, such as sarcasm detection, humour identification, and emotion recognition. Our evaluation shows that MOSES outperforms the state-of-the-art system for SED by an average of ~2% on different evaluation metrics, such as ROUGE, BLEU, and METEOR. Further, we observe that leveraging the generated explanation advances three downstream tasks for affect classification - an average improvement of ~14% F1-score in the sarcasm detection task and ~2% in the humour identification and emotion recognition task. We also perform extensive analyses to assess the quality of the results.