TY - GEN
T1 - Entropy-Based Logic Explanations of Differentiable Decision Tree
AU - Liu, Yuanyuan
AU - Zhang, Jiajia
AU - Li, Yifan
N1 - Publisher Copyright:
© IFIP International Federation for Information Processing 2024.
PY - 2024
Y1 - 2024
N2 - Explainable reinforcement learning has evolved rapidly over the years because transparency of the model’s decision-making process is crucial in some important domains. Differentiable decision trees have been applied to this field due to their performance and interpretability. However, the number of parameters per branch node of a differentiable decision tree is related to the state dimension. When the feature dimension of states increases, the number of states considered by the model in each branch node decision also increases linearly, which increases the difficulty of human understanding. This paper proposes a entroy-based differentiable decision tree, which can restrict each branch node to use as few features as possible to predict during the training process. After the training is completed, the parameters that have little impact on the output of the branch node will be blocked, thus significantly reducing the decision complexity of each branch node. Experiments in multiple environments demonstrate the significant interpretability advantage of our proposed approach.
AB - Explainable reinforcement learning has evolved rapidly over the years because transparency of the model’s decision-making process is crucial in some important domains. Differentiable decision trees have been applied to this field due to their performance and interpretability. However, the number of parameters per branch node of a differentiable decision tree is related to the state dimension. When the feature dimension of states increases, the number of states considered by the model in each branch node decision also increases linearly, which increases the difficulty of human understanding. This paper proposes a entroy-based differentiable decision tree, which can restrict each branch node to use as few features as possible to predict during the training process. After the training is completed, the parameters that have little impact on the output of the branch node will be blocked, thus significantly reducing the decision complexity of each branch node. Experiments in multiple environments demonstrate the significant interpretability advantage of our proposed approach.
KW - Differentiable decision tree
KW - Entropy regularization
KW - Interpretable reinforcement learning
UR - https://www.scopus.com/pages/publications/85190678209
U2 - 10.1007/978-3-031-57808-3_6
DO - 10.1007/978-3-031-57808-3_6
M3 - 会议稿件
AN - SCOPUS:85190678209
SN - 9783031578076
T3 - IFIP Advances in Information and Communication Technology
SP - 79
EP - 91
BT - Intelligent Information Processing XII - 13th IFIP TC 12 International Conference, IIP 2024, Proceedings
A2 - Shi, Zhongzhi
A2 - Torresen, Jim
A2 - Yang, Shengxiang
PB - Springer Science and Business Media Deutschland GmbH
T2 - 13th IFIP TC 12 International Conference on Intelligent Information Processing, IIP 2024
Y2 - 3 May 2024 through 6 May 2024
ER -