论文标题
无参数的在线测试时间改编
Parameter-free Online Test-time Adaptation
论文作者
论文摘要
培训最先进的视觉模型对于研究人员和从业者来说变得非常昂贵。为了可访问性和资源重用,重要的是要专注于将这些模型调整为各种下游方案。一个有趣且实用的范式是在线测试时间适应,据此,培训数据无法访问,没有来自测试分布的标记数据,并且适应性只能在测试时间和少数样本上进行。在本文中,我们研究了在各种现实世界中的许多预训练模型的测试时间适应方法如何票价,从而大大扩展了它们最初的评估方式。我们证明,它们仅在定义狭窄的实验设置中表现良好,而当未选择对其测试的相同情况下选择超参数时,有时会灾难性地进行灾难性。我们提出了一种特别“保守”的方法,这是由于最终会在测试时遇到的条件的固有不确定性所激发的,该方法通过Laplacian调整后的最大可能性估计(Lake)目标解决了问题。通过调整模型的输出(不是其参数),并通过有效的凹形凸口程序来解决我们的目标,我们的方法在方案中的平均准确性比现有方法的平均精度高得多,同时又要快得多,并且具有较低的存储器足迹。该代码可在https://github.com/fiveai/lame上找到。
Training state-of-the-art vision models has become prohibitively expensive for researchers and practitioners. For the sake of accessibility and resource reuse, it is important to focus on adapting these models to a variety of downstream scenarios. An interesting and practical paradigm is online test-time adaptation, according to which training data is inaccessible, no labelled data from the test distribution is available, and adaptation can only happen at test time and on a handful of samples. In this paper, we investigate how test-time adaptation methods fare for a number of pre-trained models on a variety of real-world scenarios, significantly extending the way they have been originally evaluated. We show that they perform well only in narrowly-defined experimental setups and sometimes fail catastrophically when their hyperparameters are not selected for the same scenario in which they are being tested. Motivated by the inherent uncertainty around the conditions that will ultimately be encountered at test time, we propose a particularly "conservative" approach, which addresses the problem with a Laplacian Adjusted Maximum-likelihood Estimation (LAME) objective. By adapting the model's output (not its parameters), and solving our objective with an efficient concave-convex procedure, our approach exhibits a much higher average accuracy across scenarios than existing methods, while being notably faster and have a much lower memory footprint. The code is available at https://github.com/fiveai/LAME.