While the integration of product images enhances the recommendation performance of visual-based recommender systems (VRSs), this can make the model vulnerable to adversaries that can produce noised images capable to alter the recommendation behavior. Recently, stronger and stronger adversarial attacks have emerged to raise awareness of these risks; however, effective defense methods are still an urgent open challenge. In this work, we propose "Adversarial Image Denoiser" (AiD), a novel defense method that cleans up the item images by malicious perturbations. In particular, we design a training strategy whose denoising objective is to minimize both the visual differences between clean and adversarial images and preserve the ranking performance in authentic settings. We perform experiments to evaluate the efficacy of AiD using three state-of-the-art adversarial attacks mounted against standard VRSs. Code and datasets at https://github.com/sisinflab/Denoise-to-protect-VRS.

Denoise to Protect: A Method to Robustify Visual Recommenders from Adversaries / Merra, F. A.; Anelli, V. W.; Di Noia, T.; Malitesta, D.; Mancino, A. C. M.. - (2023), pp. 1924-1928. (Intervento presentato al convegno 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2023 tenutosi a twn nel 2023) [10.1145/3539618.3591971].

Denoise to Protect: A Method to Robustify Visual Recommenders from Adversaries

Merra F. A.;Anelli V. W.;Di Noia T.;Malitesta D.;Mancino A. C. M.
2023-01-01

Abstract

While the integration of product images enhances the recommendation performance of visual-based recommender systems (VRSs), this can make the model vulnerable to adversaries that can produce noised images capable to alter the recommendation behavior. Recently, stronger and stronger adversarial attacks have emerged to raise awareness of these risks; however, effective defense methods are still an urgent open challenge. In this work, we propose "Adversarial Image Denoiser" (AiD), a novel defense method that cleans up the item images by malicious perturbations. In particular, we design a training strategy whose denoising objective is to minimize both the visual differences between clean and adversarial images and preserve the ranking performance in authentic settings. We perform experiments to evaluate the efficacy of AiD using three state-of-the-art adversarial attacks mounted against standard VRSs. Code and datasets at https://github.com/sisinflab/Denoise-to-protect-VRS.
2023
46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2023
9781450394086
Denoise to Protect: A Method to Robustify Visual Recommenders from Adversaries / Merra, F. A.; Anelli, V. W.; Di Noia, T.; Malitesta, D.; Mancino, A. C. M.. - (2023), pp. 1924-1928. (Intervento presentato al convegno 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2023 tenutosi a twn nel 2023) [10.1145/3539618.3591971].
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11589/259260
Citazioni
  • Scopus 4
  • ???jsp.display-item.citation.isi??? 3
social impact