论文标题
在基准测试中挤压激活功能:使用连续值逻辑迈向可解释的人工智能
Squashing activation functions in benchmark tests: towards eXplainable Artificial Intelligence using continuous-valued logic
论文作者
论文摘要
在过去的几年中,深层神经网络在多个任务中表现出了很好的结果,但是,仍有越来越需要解决可解释性问题以提高模型透明度,性能和安全性的问题。通过将神经网络与连续的逻辑和多标准决策工具相结合是解决此问题的最有希望的方法之一:通过这种组合,可以降低神经模型的黑盒本质,可以减少可解释的人工智能(XAI)。基于连续的逻辑神经模型使用所谓的壁板激活功能,这是满足自然不变性要求并包含整流线性单元的参数函数家族。这项工作展示了第一个测量神经网络中挤压功能的性能的基准测试。进行了三个实验以检查其可用性,并针对五种不同的网络类型进行了与最流行的激活功能进行比较。通过测量每个时期的准确性,损失和时间来确定性能。这些实验和进行的基准证明,壁板功能的使用是可能的,并且在性能上与常规激活功能相似。此外,通过实施nilpotent逻辑门来证明如何成功解决和高性能解决简单的分类任务,从而进行了进一步的实验。结果表明,由于嵌入式的nilpotent逻辑运算符和挤压功能的不同性,因此可以解决其他常用的激活功能失败的分类问题。
Over the past few years, deep neural networks have shown excellent results in multiple tasks, however, there is still an increasing need to address the problem of interpretability to improve model transparency, performance, and safety. Achieving eXplainable Artificial Intelligence (XAI) by combining neural networks with continuous logic and multi-criteria decision-making tools is one of the most promising ways to approach this problem: by this combination, the black-box nature of neural models can be reduced. The continuous logic-based neural model uses so-called Squashing activation functions, a parametric family of functions that satisfy natural invariance requirements and contain rectified linear units as a particular case. This work demonstrates the first benchmark tests that measure the performance of Squashing functions in neural networks. Three experiments were carried out to examine their usability and a comparison with the most popular activation functions was made for five different network types. The performance was determined by measuring the accuracy, loss, and time per epoch. These experiments and the conducted benchmarks have proven that the use of Squashing functions is possible and similar in performance to conventional activation functions. Moreover, a further experiment was conducted by implementing nilpotent logical gates to demonstrate how simple classification tasks can be solved successfully and with high performance. The results indicate that due to the embedded nilpotent logical operators and the differentiability of the Squashing function, it is possible to solve classification problems, where other commonly used activation functions fail.