Accelerating the Surrogate Retraining for Poisoning Attacks against Recommender Systems

Y Wu, Q Cao, S Tao, K Zhang, F Sun… - arXiv preprint arXiv …, 2024 - arxiv.org
Y Wu, Q Cao, S Tao, K Zhang, F Sun, H Shen
arXiv preprint arXiv:2408.10666, 2024arxiv.org
Recent studies have demonstrated the vulnerability of recommender systems to data
poisoning attacks, where adversaries inject carefully crafted fake user interactions into the
training data of recommenders to promote target items. Current attack methods involve
iteratively retraining a surrogate recommender on the poisoned data with the latest fake
users to optimize the attack. However, this repetitive retraining is highly time-consuming,
hindering the efficient assessment and optimization of fake users. To mitigate this …
Recent studies have demonstrated the vulnerability of recommender systems to data poisoning attacks, where adversaries inject carefully crafted fake user interactions into the training data of recommenders to promote target items. Current attack methods involve iteratively retraining a surrogate recommender on the poisoned data with the latest fake users to optimize the attack. However, this repetitive retraining is highly time-consuming, hindering the efficient assessment and optimization of fake users. To mitigate this computational bottleneck and develop a more effective attack in an affordable time, we analyze the retraining process and find that a change in the representation of one user/item will cause a cascading effect through the user-item interaction graph. Under theoretical guidance, we introduce \emph{Gradient Passing} (GP), a novel technique that explicitly passes gradients between interacted user-item pairs during backpropagation, thereby approximating the cascading effect and accelerating retraining. With just a single update, GP can achieve effects comparable to multiple original training iterations. Under the same number of retraining epochs, GP enables a closer approximation of the surrogate recommender to the victim. This more accurate approximation provides better guidance for optimizing fake users, ultimately leading to enhanced data poisoning attacks. Extensive experiments on real-world datasets demonstrate the efficiency and effectiveness of our proposed GP.
arxiv.org