User-Centric Explainability in Fintech Applications
Fintech applications such as robo-financial advisors (RAs) are complex algorithmic decision making systems, which gained prominence with their claim to democratize finance. Lack of transparency and explanations for these automated decisions leads to a trust deficit for users, limiting the potential...
Saved in:
Published in | HCI International 2021 - Posters pp. 481 - 488 |
---|---|
Main Authors | , |
Format | Book Chapter |
Language | English |
Published |
Cham
Springer International Publishing
|
Series | Communications in Computer and Information Science |
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Fintech applications such as robo-financial advisors (RAs) are complex algorithmic decision making systems, which gained prominence with their claim to democratize finance. Lack of transparency and explanations for these automated decisions leads to a trust deficit for users, limiting the potential of these applications. Our research aims to analyse the effectiveness of user-centric explanations in conveying the decision-making logic of complex algorithmic systems. Our user study tests techniques from explainable AI, varying in complexity and transparency. The quantitative aspects of our study determine the efficacy and usability of explanations and the qualitative aspects measure the effect of explanations on users and system usability. Our study finds trust and confidence of users in the system is positively correlated with comprehension and transparency provided by the presence of an explanation. There is a notable reduction in comprehension and trust between transparent white and opaque black box explanations of algorithms. This study is designed to aid policymakers and regulators in order to understand user needs which are crucial to designing better policies around algorithmic explainability for RAs. |
---|---|
ISBN: | 3030786412 9783030786410 |
ISSN: | 1865-0929 1865-0937 |
DOI: | 10.1007/978-3-030-78642-7_64 |