Abstract
Decision-making systems have been widely used in the Financial Services domain. AI is bringing both many innovations and opportunities as well as new risks linked to ethical considerations. Customer trust is at the forefront of customer retention. To build trust, there is the need to make the decision process Interpretable, Understandable, and Trustworthy for the end-user. Since products offered within the banking sector are usually of an intangible nature, customer trust perception is crucial to maintain a long-standing relationship and to ensure customer loyalty. To this end, in this paper we propose more insightful and user-friendly explanations for decisions made by AI systems in the financial domain.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Barocas, S., Hardt, M., Narayanan, A.: Fairness and machine learning (2019). https://fairmlbook.org/
Bracke, P., Datta, A., Jung, C., Sen, S.: Machine learning explainability in finance: an application to default risk analysis (816) (2019)
Chouldechova, A.: Fair prediction with disparate impact: a study of bias in recidivism prediction instruments. Big Data 5(2), 153–163 (2017)
Cohen, L., Lipton, Z.C., Mansour, Y.: Efficient candidate screening under multiple tests and implications for fairness. In: FORC. LIPIcs, vol. 156, pp. 1:1–1:20. Schloss Dagstuhl - Leibniz-Zentrum für Informatik (2020)
Congress of the United States: Fair housing act (1968)
Dwork, C., Hardt, M., Pitassi, T., Reingold, O., Zemel, R.: Fairness through awareness. In: ITCS, pp. 214–226 (2012)
Federal Reserve Board: The truth in lending act (1968)
Federal Trade Commission: Equal credit opportunity act (1974)
Hardt, M., Price, E., Srebro, N.: Equality of opportunity in supervised learning. In: NIPS, pp. 3315–3323 (2016)
High-Level Expert Group on AI: Ethics guidelines for trustworthy AI. Report, European Commission, Brussels (2019)
Kozodoi, N., Jacob, J., Lessmann, S.: Fairness in credit scoring: assessment, implementation and profit implications. arXiv preprint arXiv:2103.01907 (2021)
Lundberg, S.M., Lee, S.: A unified approach to interpreting model predictions. In: NIPS, pp. 4765–4774 (2017)
Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., Galstyan, A.: A survey on bias and fairness in machine learning (2019)
Ribeiro, M.T., Singh, S., Guestrin, C.: “Why should I trust you?”: explaining the predictions of any classifier. In: KDD, pp. 1135–1144. ACM (2016)
Sharma, R., Schommer, C., Vivarelli, N.: Building up explainability in multi-layer perceptrons for credit risk modeling. In: DSAA, pp. 761–762. IEEE (2020)
Stepin, I., Alonso, J.M., Catala, A., Pereira-Fariña, M.: A survey of contrastive and counterfactual explanation generation methods for explainable artificial intelligence. IEEE Access 9, 11974–12001 (2021)
Verma, S., Rubin, J.: Fairness definitions explained. In: ICSE-FairWare, pp. 1–7 (2018)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 IFIP International Federation for Information Processing
About this paper
Cite this paper
Cornacchia, G., Narducci, F., Ragone, A. (2021). Improving the User Experience and the Trustworthiness of Financial Services. In: Ardito, C., et al. Human-Computer Interaction – INTERACT 2021. INTERACT 2021. Lecture Notes in Computer Science(), vol 12936. Springer, Cham. https://doi.org/10.1007/978-3-030-85607-6_19
Download citation
DOI: https://doi.org/10.1007/978-3-030-85607-6_19
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-85606-9
Online ISBN: 978-3-030-85607-6
eBook Packages: Computer ScienceComputer Science (R0)