论文标题
可再现模型蒸馏的通用方法
A Generic Approach for Reproducible Model Distillation
论文作者
论文摘要
型号蒸馏一直是生产可解释的机器学习的流行方法。它使用可解释的“学生”模型来模仿黑匣子“老师”模型的预测。但是,当学生模型对用于培训的数据集的可变性敏感时,即使保持教师固定,相应的解释也不可靠。现有策略通过检查是否生成了足够大的伪data语料库来可靠地重现学生模型,从而稳定模型蒸馏,但是到目前为止,已经为特定的学生模型开发了这样做的方法。在本文中,我们开发了一种基于中央限制定理的平均损失的稳定模型蒸馏的通用方法。我们从候选学生模型的集合开始,并搜索与老师合理同意的候选人。然后,我们构建一个多个测试框架来选择语料库大小,以便将一致的学生模型在不同的伪样本下选择。我们证明了我们所提出的方法在三个常用的可理解模型上的应用:决策树,下降规则列表和符号回归。最后,我们对乳腺X线量和乳腺癌数据集进行了模拟实验,并在Markov过程的理论分析中说明了测试程序。该代码可在https://github.com/yunzhe-zhou/genericdistillation上公开获取。
Model distillation has been a popular method for producing interpretable machine learning. It uses an interpretable "student" model to mimic the predictions made by the black box "teacher" model. However, when the student model is sensitive to the variability of the data sets used for training even when keeping the teacher fixed, the corresponded interpretation is not reliable. Existing strategies stabilize model distillation by checking whether a large enough corpus of pseudo-data is generated to reliably reproduce student models, but methods to do so have so far been developed for a specific student model. In this paper, we develop a generic approach for stable model distillation based on central limit theorem for the average loss. We start with a collection of candidate student models and search for candidates that reasonably agree with the teacher. Then we construct a multiple testing framework to select a corpus size such that the consistent student model would be selected under different pseudo samples. We demonstrate the application of our proposed approach on three commonly used intelligible models: decision trees, falling rule lists and symbolic regression. Finally, we conduct simulation experiments on Mammographic Mass and Breast Cancer datasets and illustrate the testing procedure throughout a theoretical analysis with Markov process. The code is publicly available at https://github.com/yunzhe-zhou/GenericDistillation.