论文标题

循环中的决策树的互动加强学习特征选择的功能选择

Interactive Reinforcement Learning for Feature Selection with Decision Tree in the Loop

论文作者

Fan, Wei, Liu, Kunpeng, Liu, Hao, Ge, Yong, Xiong, Hui, Fu, Yanjie

论文摘要

我们研究了自动化功能选择中平衡有效性和效率的问题。在探索了许多特征选择方法之后,我们观察到一个计算困境:1)传统特征选择大多有效,但很难识别最佳子集; 2)新出现的增强功能选择会自动导航到最佳子集,但通常效率低下。我们可以弥合自动化下有效性与效率之间的差距吗?在这种困境中,我们旨在开发一种新颖的功能空间导航方法。在我们的初步工作中,我们利用互动式增强学习来加速外部培训师互动的特征选择。在此期刊版本中,我们提出了一种新颖的交互式和闭环体系结构,以同时建模交互式增强学习(IRL)和决策树反馈(DTF)。具体而言,IRL将创建一个交互式特征选择循环,而DTF是将结构化的特征知识馈回循环。首先,利用决策树的树结构特征层次结构来改善状态表示。特别是,我们将所选的特征子集表示为特征功能相关性和有向特征树的无向图。我们提出了一种新的嵌入方法,能够授权图形卷积网络从图和树中共同学习状态表示。其次,利用树木结构的特征层次结构来开发新的奖励方案。特别是,我们根据决策树特征的重要性来个性化代理的奖励分配。此外,观察代理的行为可以是反馈,我们设计了另一种奖励方案,以根据历史行动记录中选定的频率比率进行权衡和分配奖励。最后,我们在现实世界数据集上介绍了广泛的实验,以显示出改进的性能。

We study the problem of balancing effectiveness and efficiency in automated feature selection. After exploring many feature selection methods, we observe a computational dilemma: 1) traditional feature selection is mostly efficient, but difficult to identify the best subset; 2) the emerging reinforced feature selection automatically navigates to the best subset, but is usually inefficient. Can we bridge the gap between effectiveness and efficiency under automation? Motivated by this dilemma, we aim to develop a novel feature space navigation method. In our preliminary work, we leveraged interactive reinforcement learning to accelerate feature selection by external trainer-agent interaction. In this journal version, we propose a novel interactive and closed-loop architecture to simultaneously model interactive reinforcement learning (IRL) and decision tree feedback (DTF). Specifically, IRL is to create an interactive feature selection loop and DTF is to feed structured feature knowledge back to the loop. First, the tree-structured feature hierarchy from decision tree is leveraged to improve state representation. In particular, we represent the selected feature subset as an undirected graph of feature-feature correlations and a directed tree of decision features. We propose a new embedding method capable of empowering graph convolutional network to jointly learn state representation from both the graph and the tree. Second, the tree-structured feature hierarchy is exploited to develop a new reward scheme. In particular, we personalize reward assignment of agents based on decision tree feature importance. In addition, observing agents' actions can be feedback, we devise another reward scheme, to weigh and assign reward based on the feature selected frequency ratio in historical action records. Finally, we present extensive experiments on real-world datasets to show the improved performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源