论文标题
基于句子结构分析使用机器学习方法的自动问题生成
Automatic question generation based on sentence structure analysis using machine learning approach
论文作者
论文摘要
自动问题生成是自然语言处理中最具挑战性的任务之一。它需要“双向”语言处理:首先,系统必须了解输入文本(自然语言理解),然后还必须以文本(自然语言生成)的形式生成问题。在本文中,我们介绍了我们的框架,以用英语从非结构化文本中产生事实问题。它结合了基于句子模式的传统语言方法以及几种机器学习方法。我们首先从输入文本中获得词汇,句法和语义信息,然后为每个句子构造一组层次的模式。从模式中提取了一组功能,然后将其用于自动学习新的转换规则。我们的学习过程完全由数据驱动,因为转换规则是从一组初始句子问题对获得的。这种方法的优点在于新的转型规则的简单扩展,这使我们能够通过增强学习来产生各种类型的问题,也可以持续改进系统。该框架还包括一个问题评估模块,该模块估计了生成问题的质量。它是选择最佳问题并消除错误或重复的过滤器。我们已经进行了几项实验来评估生成问题的正确性,并且我们还将系统与几个最新系统进行了比较。我们的结果表明,生成的问题的质量的表现优于最先进的系统,我们的问题也与人类提出的问题相媲美。我们还与所有创建的数据集并评估了问题,并发布了一个界面,因此可以跟进我们的工作。
Automatic question generation is one of the most challenging tasks of Natural Language Processing. It requires "bidirectional" language processing: firstly, the system has to understand the input text (Natural Language Understanding) and it then has to generate questions also in the form of text (Natural Language Generation). In this article, we introduce our framework for generating the factual questions from unstructured text in the English language. It uses a combination of traditional linguistic approaches based on sentence patterns with several machine learning methods. We firstly obtain lexical, syntactic and semantic information from an input text and we then construct a hierarchical set of patterns for each sentence. The set of features is extracted from the patterns and it is then used for automated learning of new transformation rules. Our learning process is totally data-driven because the transformation rules are obtained from a set of initial sentence-question pairs. The advantages of this approach lie in a simple expansion of new transformation rules which allows us to generate various types of questions and also in the continuous improvement of the system by reinforcement learning. The framework also includes a question evaluation module which estimates the quality of generated questions. It serves as a filter for selecting the best questions and eliminating incorrect ones or duplicates. We have performed several experiments to evaluate the correctness of generated questions and we have also compared our system with several state-of-the-art systems. Our results indicate that the quality of generated questions outperforms the state-of-the-art systems and our questions are also comparable to questions created by humans. We have also created and published an interface with all created datasets and evaluated questions, so it is possible to follow up on our work.