Trust and Reliance in XAI -- Distinguishing Between Attitudinal and Behavioral Measures
Trust is often cited as an essential criterion for the effective use and real-world deployment of AI. Researchers argue that AI should be more transparent to increase trust, making transparency one of the main goals of XAI. Nevertheless, empirical research on this topic is inconclusive regarding the...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Trust is often cited as an essential criterion for the effective use and
real-world deployment of AI. Researchers argue that AI should be more
transparent to increase trust, making transparency one of the main goals of
XAI. Nevertheless, empirical research on this topic is inconclusive regarding
the effect of transparency on trust. An explanation for this ambiguity could be
that trust is operationalized differently within XAI. In this position paper,
we advocate for a clear distinction between behavioral (objective) measures of
reliance and attitudinal (subjective) measures of trust. However, researchers
sometimes appear to use behavioral measures when intending to capture trust,
although attitudinal measures would be more appropriate. Based on past
research, we emphasize that there are sound theoretical reasons to keep trust
and reliance separate. Properly distinguishing these two concepts provides a
more comprehensive understanding of how transparency affects trust and
reliance, benefiting future XAI research. |
---|---|
DOI: | 10.48550/arxiv.2203.12318 |