论文标题

测试学习程序是否校准

Testing whether a Learning Procedure is Calibrated

论文作者

Cockayne, Jon, Graham, Matthew M., Oates, Chris J., Sullivan, T. J., Teymur, Onur

论文摘要

学习过程将数据集作为输入,并对假定引起数据集的模型的参数进行推断。在这里,我们考虑其输出是概率分布的学习过程,在查看数据集后代表大约$θ$的不确定性。贝叶斯推断是这样一个过程的一个典型示例,但也可以构建其他学习过程以返回分配输出。本文研究了要考虑校准的学习过程的条件,从某种意义上说,真正的数据生成参数是从其分布输出中作为样本的合理的。一个学习过程的推论和预测是系统地过度或不足的,将无法校准。另一方面,校准的学习过程不必在统计上有效。开发了一个假设测试框架,以便使用模拟评估学习程序是否经过校准。提出了几个小插图,以说明框架的不同方面。

A learning procedure takes as input a dataset and performs inference for the parameters $θ$ of a model that is assumed to have given rise to the dataset. Here we consider learning procedures whose output is a probability distribution, representing uncertainty about $θ$ after seeing the dataset. Bayesian inference is a prime example of such a procedure, but one can also construct other learning procedures that return distributional output. This paper studies conditions for a learning procedure to be considered calibrated, in the sense that the true data-generating parameters are plausible as samples from its distributional output. A learning procedure whose inferences and predictions are systematically over- or under-confident will fail to be calibrated. On the other hand, a learning procedure that is calibrated need not be statistically efficient. A hypothesis-testing framework is developed in order to assess, using simulation, whether a learning procedure is calibrated. Several vignettes are presented to illustrate different aspects of the framework.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源