论文标题
编码贝叶斯神经网络的潜在后部以进行不确定性定量
Encoding the latent posterior of Bayesian Neural Networks for uncertainty quantification
论文作者
论文摘要
长期以来,贝叶斯神经网络(BNN)被认为是改善深神经网络的鲁棒性和预测性不确定性的理想但不足的解决方案。尽管他们可以更准确地捕获网络参数的后验分布,但大多数BNN方法要么限于小型网络,要么依赖于约束假设,例如参数独立性。这些缺点使简单但在计算上的沉重方法(例如深层合奏)的突出性,其训练和测试成本随着网络数量线性增加。在这项工作中,我们旨在有效地适合复杂的计算机视觉体系结构,例如Resnet50 DeepLabV3+和任务,例如语义分割,对参数的假设较少。我们通过利用变异自动编码器(VAE)来学习每个网络层的参数的相互作用和潜在分布来实现这一目标。我们的方法是潜在的BNN(LP-BNN),与最近的batchensemble方法兼容,从而导致高效({在训练和测试期间的计算和}内存)合奏。 LP-BNN在几个具有挑战性的基准测试中获得了多个指标的竞争结果,用于图像分类,语义分割和分布外检测。
Bayesian neural networks (BNNs) have been long considered an ideal, yet unscalable solution for improving the robustness and the predictive uncertainty of deep neural networks. While they could capture more accurately the posterior distribution of the network parameters, most BNN approaches are either limited to small networks or rely on constraining assumptions such as parameter independence. These drawbacks have enabled prominence of simple, but computationally heavy approaches such as Deep Ensembles, whose training and testing costs increase linearly with the number of networks. In this work we aim for efficient deep BNNs amenable to complex computer vision architectures, e.g. ResNet50 DeepLabV3+, and tasks, e.g. semantic segmentation, with fewer assumptions on the parameters. We achieve this by leveraging variational autoencoders (VAEs) to learn the interaction and the latent distribution of the parameters at each network layer. Our approach, Latent-Posterior BNN (LP-BNN), is compatible with the recent BatchEnsemble method, leading to highly efficient ({in terms of computation and} memory during both training and testing) ensembles. LP-BNN s attain competitive results across multiple metrics in several challenging benchmarks for image classification, semantic segmentation and out-of-distribution detection.