论文标题
自动:通过多样性感知的互动加强学习选择自动化功能
AutoFS: Automated Feature Selection via Diversity-aware Interactive Reinforcement Learning
论文作者
论文摘要
在本文中,我们研究了自动化特征选择中平衡有效性和效率的问题。特征选择是用于机器学习和预测分析的基本智能。在探索了许多特征选择方法之后,我们观察到计算困境:1)传统特征选择方法(例如MRMR)大多是有效的,但很难识别最佳的子集; 2)新兴的增强功能选择方法自动导航功能空间以探索最佳子集,但通常效率低下。自动化和效率总是彼此不同吗?我们可以弥合自动化下有效性与效率之间的差距吗?这项研究是由这种计算困境的动机,是为了开发一种新型的特征空间导航方法。为此,我们提出了一个交互式增强功能选择(IRFS)框架,该框架不仅通过自我探索的经验来指导代理人,而且还通过多样化的外部熟练教练来加速学习以进行功能探索。具体而言,我们将特征选择问题提出为交互式增强学习框架。在此框架中,我们首先对两个熟练的搜索策略熟练的培训师建模:(1)基于Kbest的培训师; (2)基于决策树的教练。然后,我们制定了两种策略:(1)确定自信和犹豫的代理人以使代理人培训多样化,(2)使两位教练能够在不同阶段扮演教学角色,以融合教练的经验并多样化教学过程。这种混合教学策略可以帮助代理商学习更广泛的知识,然后更有效。最后,我们在现实世界数据集上介绍了广泛的实验,以证明我们方法的改进性能:比现有的增强选择更有效,比经典选择更有效。
In this paper, we study the problem of balancing effectiveness and efficiency in automated feature selection. Feature selection is a fundamental intelligence for machine learning and predictive analysis. After exploring many feature selection methods, we observe a computational dilemma: 1) traditional feature selection methods (e.g., mRMR) are mostly efficient, but difficult to identify the best subset; 2) the emerging reinforced feature selection methods automatically navigate feature space to explore the best subset, but are usually inefficient. Are automation and efficiency always apart from each other? Can we bridge the gap between effectiveness and efficiency under automation? Motivated by such a computational dilemma, this study is to develop a novel feature space navigation method. To that end, we propose an Interactive Reinforced Feature Selection (IRFS) framework that guides agents by not just self-exploration experience, but also diverse external skilled trainers to accelerate learning for feature exploration. Specifically, we formulate the feature selection problem into an interactive reinforcement learning framework. In this framework, we first model two trainers skilled at different searching strategies: (1) KBest based trainer; (2) Decision Tree based trainer. We then develop two strategies: (1) to identify assertive and hesitant agents to diversify agent training, and (2) to enable the two trainers to take the teaching role in different stages to fuse the experiences of the trainers and diversify teaching process. Such a hybrid teaching strategy can help agents to learn broader knowledge, and, thereafter, be more effective. Finally, we present extensive experiments on real-world datasets to demonstrate the improved performances of our method: more efficient than existing reinforced selection and more effective than classic selection.