论文标题

提示:黑框文本分类,带有十个正向通行证

PromptBoosting: Black-Box Text Classification with Ten Forward Passes

论文作者

Hou, Bairu, O'Connor, Joe, Andreas, Jacob, Chang, Shiyu, Zhang, Yang

论文摘要

我们描述了提示启动,这是一种从神经语言模型(LM)构建文本分类器的查询过程,而无需访问LM的参数,梯度或隐藏表示形式。随着大规模LMS的培训成本和推断,这种“黑盒”分类器培训形式变得越来越重要。但是,现有的黑框LM分类器学习方法本身在计算上效率低下,通常通过使用Zeroth-rorder优化方法在大的(离散或连续)提示的大空间中进行搜索,通常将LMS专用于目标任务。 ProfteBoosting没有直接在及时空间中进行优化,而是通过无坡度的方法获得了一小部分提示,然后通过将这些提示与LM输出分布的不同元素配对来构建大量的弱学习者。然后,使用Adaboost算法结合这些弱学习者。整个学习过程只需要少量的远期通行证,而无需向后传球。实验表明,迅速启动在多个黑盒子几个少量分类任务中实现最先进的表现,并且在几次和标准学习范式中匹配或胜过全面的微调,而训练比现有的Black-box方法快10倍。

We describe PromptBoosting, a query-efficient procedure for building a text classifier from a neural language model (LM) without access to the LM's parameters, gradients, or hidden representations. This form of "black-box" classifier training has become increasingly important as the cost of training and inference in large-scale LMs grows. But existing black-box LM classifier learning approaches are themselves computationally inefficient, typically specializing LMs to the target task by searching in a large space of (discrete or continuous) prompts using zeroth-order optimization methods. Instead of directly optimizing in prompt space, PromptBoosting obtains a small pool of prompts via a gradient-free approach and then constructs a large pool of weak learners by pairing these prompts with different elements of the LM's output distribution. These weak learners are then ensembled using the AdaBoost algorithm. The entire learning process requires only a small number of forward passes and no backward pass. Experiments show that PromptBoosting achieves state-of-the-art performance in multiple black-box few-shot classification tasks, and matches or outperforms full fine-tuning in both few-shot and standard learning paradigms, while training 10x faster than existing black-box methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源