Deep learning-based recommender systems have become an integral part of several online platforms. However, their black-box nature emphasizes the need for explainable artificial intelligence (XAI) approaches to provide human-understandable reasons why a specific item gets recommended to a given user. One such method is counterfactual explanation (CF). While CFs can be highly beneficial for users and system designers, malicious actors may also exploit these explanations to undermine the system's security. In this work, we propose H-CARS, a novel strategy to poison recommender systems via CFs. Specifically, we first train a logical-reasoning-based surrogate model on training data derived from counterfactual explanations. By reversing the learning process of the recommendation model, we thus develop a proficient greedy algorithm to generate fabricated user profiles and their associated interaction records for the aforementioned surrogate model. Our experiments, which employ a well-known CF generation method and are conducted on two distinct datasets, show that H-CARS yields significant and successful attack performance.

The Dark Side of Explanations: Poisoning Recommender Systems with Counterfactual Examples / Chen, Ziheng; Silvestri, Fabrizio; Wang, Jia; Zhang, Yongfeng; Tolomei, Gabriele. - (2023), pp. 2426-2430. ( 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2023 Taipei, Taiwan ) [10.1145/3539618.3592070].

The Dark Side of Explanations: Poisoning Recommender Systems with Counterfactual Examples

Silvestri, Fabrizio
;
Tolomei, Gabriele
2023

Abstract

Deep learning-based recommender systems have become an integral part of several online platforms. However, their black-box nature emphasizes the need for explainable artificial intelligence (XAI) approaches to provide human-understandable reasons why a specific item gets recommended to a given user. One such method is counterfactual explanation (CF). While CFs can be highly beneficial for users and system designers, malicious actors may also exploit these explanations to undermine the system's security. In this work, we propose H-CARS, a novel strategy to poison recommender systems via CFs. Specifically, we first train a logical-reasoning-based surrogate model on training data derived from counterfactual explanations. By reversing the learning process of the recommendation model, we thus develop a proficient greedy algorithm to generate fabricated user profiles and their associated interaction records for the aforementioned surrogate model. Our experiments, which employ a well-known CF generation method and are conducted on two distinct datasets, show that H-CARS yields significant and successful attack performance.
2023
46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2023
explainable recommender systems; counterfactual explanations; model poisoning attacks
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
The Dark Side of Explanations: Poisoning Recommender Systems with Counterfactual Examples / Chen, Ziheng; Silvestri, Fabrizio; Wang, Jia; Zhang, Yongfeng; Tolomei, Gabriele. - (2023), pp. 2426-2430. ( 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2023 Taipei, Taiwan ) [10.1145/3539618.3592070].
File allegati a questo prodotto
File Dimensione Formato  
Chen_The-Dark-Side_2023.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.17 MB
Formato Adobe PDF
1.17 MB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1686336
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 15
  • ???jsp.display-item.citation.isi??? 6
social impact