论文标题
自然语言的幻觉调查
Survey of Hallucination in Natural Language Generation
论文作者
论文摘要
由于序列到序列深度学习技术(例如基于变压器的语言模型)的发展,自然语言产生(NLG)在近年来取得了成倍的改善。这一进步导致了更加流利和连贯的NLG,从而改善了下游任务的发展,例如抽象摘要,对话生成和数据对文本生成。但是,很明显,基于深度学习的一代容易幻觉意想不到的文本,这会降低系统性能,并且在许多真实世界的情况下无法满足用户的期望。为了解决这个问题,在测量和减轻幻觉文本方面已经介绍了许多研究,但以前从未以全面的方式进行审查。因此,在这项调查中,我们对NLG幻觉问题的研究进度和挑战提供了广泛的概述。该调查分为两个部分:(1)指标,缓解方法和未来方向的一般概述; (2)在以下下游任务中,幻觉的特定任务研究进度概述,即抽象性摘要,对话生成,生成问题回答,数据对文本生成,机器翻译和视觉语言生成; (3)大语言模型(LLMS)中的幻觉。这项调查旨在促进研究人员之间在NLG中应对幻觉文本的挑战方面的合作努力。
Natural Language Generation (NLG) has improved exponentially in recent years thanks to the development of sequence-to-sequence deep learning technologies such as Transformer-based language models. This advancement has led to more fluent and coherent NLG, leading to improved development in downstream tasks such as abstractive summarization, dialogue generation and data-to-text generation. However, it is also apparent that deep learning based generation is prone to hallucinate unintended text, which degrades the system performance and fails to meet user expectations in many real-world scenarios. To address this issue, many studies have been presented in measuring and mitigating hallucinated texts, but these have never been reviewed in a comprehensive manner before. In this survey, we thus provide a broad overview of the research progress and challenges in the hallucination problem in NLG. The survey is organized into two parts: (1) a general overview of metrics, mitigation methods, and future directions; (2) an overview of task-specific research progress on hallucinations in the following downstream tasks, namely abstractive summarization, dialogue generation, generative question answering, data-to-text generation, machine translation, and visual-language generation; and (3) hallucinations in large language models (LLMs). This survey serves to facilitate collaborative efforts among researchers in tackling the challenge of hallucinated texts in NLG.