Knowledge-aware Interpretable Recommender Systems – Knowledge Graphs for eXplainable AI
AuthorsAnelli Vito Walter
, Bellini Vito, Di Noia Tommaso
, Di Sciascio Eugenio
Recommender systems are everywhere, from e-commerce to streaming platforms. They help users lost in the maze of available information, items and services to find their way. Among them, over the years, approaches based on machine learning techniques have shown particularly good performance for top-N recommendations engines. Unfortunately, they mostly behave as black-boxes and, even when they embed some form of description about the items to recommend, after the training phase they move such descriptions in a latent space thus loosing the actual explicit semantics of recommended items. As a consequence, the system designers struggle at providing satisfying explanations to the recommendation list provided to the end user. In this chapter, we describe two approaches to recommendation which make use of the semantics encoded in a knowledge graph to train interpretable models which keep the original semantics of the items description thus providing a powerful tool to automatically compute explainable results. The two methods relies on two completely different machine learning algorithms, namely, factorization machines and autoencoder neural networks. We also show how to measure the interpretability of the model through the introduction of two metrics: semantic accuracy and robustness.
author = "Anelli, Vito Walter and Bellini, Vito and Di Noia, Tommaso and Di Sciascio, Eugenio",
title = "Knowledge-aware Interpretable Recommender Systems - Knowledge Graphs for eXplainable AI",
series = "Foundations, Applications and Challenges. Studies on the Semantic Web",
volume = "47",
pages = "101--124",
year = "2020",
publisher = "IOS Press, Amsterdam",
note = "doi:10.3233/SSW200014 isbn: 978-1-64368-080-4",
url = "http://sisinflab.poliba.it/Publications/2020/ABDD20"