Shallow decision trees for explainable k-means clustering
•First explainable clustering algorithm for k-means sensitive to the tree’s depth.•Superior results to similar algorithms in experiments with 16 datasets.•Parameter calibration allows exploration of trade-off between cost and explainability. A number of recent works have employed decision trees for...
Gespeichert in:
Veröffentlicht in: | Pattern recognition 2023-05, Vol.137, p.109239, Article 109239 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | •First explainable clustering algorithm for k-means sensitive to the tree’s depth.•Superior results to similar algorithms in experiments with 16 datasets.•Parameter calibration allows exploration of trade-off between cost and explainability.
A number of recent works have employed decision trees for the construction of explainable partitions that aim to minimize the k-means cost function. These works, however, largely ignore metrics related to the depths of the leaves in the resulting tree, which is perhaps surprising considering how the explainability of a decision tree depends on these depths. To fill this gap in the literature, we propose an efficient algorithm with a penalty term in its loss function to favor the construction of shallow decision trees – i.e., trees whose leaves are not very deep, which translate to clusters that are defined by a small number of attributes and are therefore easier to explain. In experiments on 16 datasets, our algorithm yields better results than decision-tree clustering algorithms recently presented in the literature, typically achieving lower or equivalent costs with considerably shallower trees. |
---|---|
ISSN: | 0031-3203 1873-5142 |
DOI: | 10.1016/j.patcog.2022.109239 |