Differential Privacy in Collaborative Filtering Recommender Systems: A Review

Abstract

State-of-the-art recommender systems produce high-quality recommendations to support users in finding relevant content. However, through the utilization of users' data for generating recommendations, recommender systems threaten users' privacy. To alleviate this threat, often, differential privacy is used to protect users' data via adding random noise. This, however, leads to a substantial drop in recommendation quality. Therefore, several approaches aim to improve this trade-off between accuracy and user privacy. In this work, we first overview threats to user privacy in recommender systems, followed by a brief introduction to the differential privacy framework that can protect users' privacy. Subsequently, we review recommendation approaches that apply differential privacy, and we highlight research that improves the trade-off between recommendation quality and user privacy. Finally, we discuss open issues, e.g., considering the relation between privacy and fairness, and the users' different needs for privacy. With this review, we hope to provide other researchers an overview of the ways in which differential privacy has been applied to state-of-the-art collaborative filtering recommender systems.


Citation

Peter Müllner, Elisabeth Lex, Markus Schedl, Dominik Kowald
Differential Privacy in Collaborative Filtering Recommender Systems: A Review
Frontiers in Big Data - Recommender Systems, 6: doi:10.3389/fdata.2023.1249997, 2023.

BibTeX

@article{Müllner2023DiffPrivacy,
    title = {Differential Privacy in Collaborative Filtering Recommender Systems: A Review},
    author = {Müllner, Peter and Lex, Elisabeth and Schedl, Markus and Kowald, Dominik},
    journal = {Frontiers in Big Data - Recommender Systems},
    doi = {10.3389/fdata.2023.1249997},
    url = {https://doi.org/10.3389/fdata.2023.1249997},
    volume = {6},
    year = {2023}
}