论文标题
使用整数加权条款扩展TSETLIN机器,以提高可解释性
Extending the Tsetlin Machine With Integer-Weighted Clauses for Increased Interpretability
论文作者
论文摘要
尽管付出了巨大的努力,但对于许多模式识别问题来说,既可以解释又准确的建筑模型是尚未解决的挑战。通常,基于规则的线性模型缺乏准确性,而深度学习的解释性基于基础推断的粗略近似。 Tsetlin机器(TMS)使用结合子句的线性组合,在不同的基准测试中表现出竞争性能。但是,为此,需要许多子句,这会影响解释性。在这里,我们通过为TM条款提供整数权重来解决机器学习中的准确性解干性挑战。由此产生的整数加权TM(IWTM)涉及学习哪些条款不准确的问题,因此必须组合以获得高精度作为团队(低权重子句),以及哪些条款足够准确以更加独立地运作(高权重子句)。由于每个TM子句都是由Tsetlin Automata团队自适应形成的,因此确定有效的权重成为一个具有挑战性的在线学习问题。我们通过在线路(SSL)自动机上进行随机搜索来扩展Tsetlin Automata的每个团队来解决此问题。在我们的新颖方案中,SSL自动机通过与相应的Tsetlin Automata团队进行互动,了解其子句的重量,进而通过调整权重调整该子句的组成。我们使用五个数据集对IWTM进行了经验评估,包括互相性研究。平均而言,IWTM使用的文字比香草TM少的6.5倍,而文字少120倍,其文字比具有实值重量的TM少120倍。此外,就平均F1分数而言,IWTM的表现优于简单的多层人工神经网络,决策树,支持向量机,K-Nearest邻居,随机森林,随机森林,XGBoost,可解释的增强机器以及标准和真实价值的加权TMS。
Despite significant effort, building models that are both interpretable and accurate is an unresolved challenge for many pattern recognition problems. In general, rule-based and linear models lack accuracy, while deep learning interpretability is based on rough approximations of the underlying inference. Using a linear combination of conjunctive clauses in propositional logic, Tsetlin Machines (TMs) have shown competitive performance on diverse benchmarks. However, to do so, many clauses are needed, which impacts interpretability. Here, we address the accuracy-interpretability challenge in machine learning by equipping the TM clauses with integer weights. The resulting Integer Weighted TM (IWTM) deals with the problem of learning which clauses are inaccurate and thus must team up to obtain high accuracy as a team (low weight clauses), and which clauses are sufficiently accurate to operate more independently (high weight clauses). Since each TM clause is formed adaptively by a team of Tsetlin Automata, identifying effective weights becomes a challenging online learning problem. We address this problem by extending each team of Tsetlin Automata with a stochastic searching on the line (SSL) automaton. In our novel scheme, the SSL automaton learns the weight of its clause in interaction with the corresponding Tsetlin Automata team, which, in turn, adapts the composition of the clause by the adjusting weight. We evaluate IWTM empirically using five datasets, including a study of interpetability. On average, IWTM uses 6.5 times fewer literals than the vanilla TM and 120 times fewer literals than a TM with real-valued weights. Furthermore, in terms of average F1-Score, IWTM outperforms simple Multi-Layered Artificial Neural Networks, Decision Trees, Support Vector Machines, K-Nearest Neighbor, Random Forest, XGBoost, Explainable Boosting Machines, and standard and real-value weighted TMs.