skip to main content
research-article

Semantic-Aware Adversarial Training for Reliable Deep Hashing Retrieval

Published: 01 January 2023 Publication History

Abstract

Deep hashing has been intensively studied and successfully applied in large-scale image retrieval systems due to its efficiency and effectiveness. Recent studies have recognized that the existence of adversarial examples poses a security threat to deep hashing models, that is, adversarial vulnerability. Notably, it is challenging to efficiently distill reliable semantic representatives for deep hashing to guide adversarial learning, and thereby it hinders the enhancement of adversarial robustness of deep hashing-based retrieval models. Moreover, current researches on adversarial training for deep hashing are hard to be formalized into a unified minimax structure. In this paper, we explore Semantic-Aware Adversarial Training (SAAT) for improving the adversarial robustness of deep hashing models. Specifically, we conceive a discriminative mainstay features learning (DMFL) scheme to construct semantic representatives for guiding adversarial learning in deep hashing. Particularly, our DMFL with the strict theoretical guarantee is adaptively optimized in a discriminative learning manner, where both discriminative and semantic properties are jointly considered. Moreover, adversarial examples are fabricated by maximizing the Hamming distance between the hash codes of adversarial samples and mainstay features, the efficacy of which is validated in the adversarial attack trials. Further, we, for the first time, formulate the formalized adversarial training of deep hashing into a unified minimax optimization under the guidance of the generated mainstay codes. Extensive experiments on benchmark datasets show superb attack performance against the state-of-the-art algorithms, meanwhile, the proposed adversarial training can effectively eliminate adversarial perturbations for trustworthy deep hashing-based retrieval. Our code is available at <uri>https://github.com/xandery-geek/SAAT</uri>.

References

[1]
A. Andoni and P. Indyk, “Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions,” in Proc. 47th Annu. IEEE Symp. Found. Comput. Sci. (FOCS), 2006, pp. 459–468.
[2]
J. Wang, T. Zhang, J. Song, N. Sebe, and H. T. Shen, “A survey on learning to hash,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 40, no. 4, pp. 769–790, Apr. 2018.
[3]
R. Xia, Y. Pan, H. Lai, C. Liu, and S. Yan, “Supervised hashing for image retrieval via image representation learning,” in Proc. AAAI Conf. Artif. Intell., 2014, pp. 2156–2162.
[4]
W. Li, S. Wang, and W. Kang, “Feature learning based deep supervised hashing with pairwise labels,” in Proc. Int. Joint Conf. Artif. Intell., 2016, pp. 1711–1717.
[5]
Z. Cao, M. Long, J. Wang, and P. S. Yu, “HashNet: Deep learning to hash by continuation,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), Oct. 2017, pp. 5609–5618.
[6]
V. Talreja, M. C. Valenti, and N. M. Nasrabadi, “Deep hashing for secure multimodal biometrics,” IEEE Trans. Inf. Forensics Security, vol. 16, pp. 1306–1321, 2021.
[7]
L. Yuanet al., “Central similarity quantization for efficient image and video retrieval,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2020, pp. 3080–3089.
[8]
Y. Wang, X. Ou, J. Liang, and Z. Sun, “Deep semantic reconstruction hashing for similarity retrieval,” IEEE Trans. Circuits Syst. Video Technol., vol. 31, no. 1, pp. 387–400, Jan. 2021.
[9]
L. Fan, K. W. Ng, C. Ju, T. Zhang, and C. S. Chan, “Deep polarized network for supervised learning of accurate binary hashing codes,” in Proc. 29th Int. Joint Conf. Artif. Intell., Jul. 2020, pp. 825–831.
[10]
J. T. Hoe, K. W. Ng, T. Zhang, C. S. Chan, Y.-Z. Song, and T. Xiang, “One loss for all: Deep hashing with a single cosine similarity based learning objective,” in Proc. Adv. Neural Inf. Process. Syst., vol. 34. Red Hook, NY, USA: Curran Associates, 2021, pp. 24286–24298.
[11]
K. D. Doan, P. Yang, and P. Li, “One loss for quantization: Deep hashing with discrete Wasserstein distributional matching,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2022, pp. 9437–9447.
[12]
D. Wu, Q. Dai, B. Li, and W. Wang, “Deep uncoupled discrete hashing via similarity matrix decomposition,” ACM Trans. Multimedia Comput., Commun., Appl., vol. 19, no. 1, pp. 1–22, Jan. 2023.
[13]
E. Yang, T. Liu, C. Deng, and D. Tao, “Adversarial examples for Hamming space search,” IEEE Trans. Cybern., vol. 50, no. 4, pp. 1473–1484, Apr. 2020.
[14]
J. Baiet al., “Targeted attack for deep hashing based retrieval,” in Proc. Eur. Conf. Comput. Vis., Aug. 2020, pp. 618–634.
[15]
X. Wang, Z. Zhang, B. Wu, F. Shen, and G. Lu, “Prototype-supervised adversarial network for targeted attack of deep hashing,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2021, pp. 16352–16361.
[16]
Z. Zhang, X. Wang, G. Lu, F. Shen, and L. Zhu, “Targeted attack of deep hashing via prototype-supervised adversarial networks,” IEEE Trans. Multimedia, vol. 24, pp. 3392–3404, 2022.
[17]
J. Lu, M. Chen, Y. Sun, W. Wang, Y. Wang, and X. Yang, “A smart adversarial attack on deep hashing based image retrieval,” in Proc. Int. Conf. Multimedia Retr., Aug. 2021, pp. 227–235.
[18]
C. Szegedyet al., “Intriguing properties of neural networks,” in Proc. Int. Conf. Learn. Represent., 2014, pp. 1–10.
[19]
L. Amsaleget al., “High intrinsic dimensionality facilitates adversarial attack: Theoretical evidence,” IEEE Trans. Inf. Forensics Security, vol. 16, pp. 854–865, 2021.
[20]
Y. Zhong and W. Deng, “Towards transferable adversarial attack against deep face recognition,” IEEE Trans. Inf. Forensics Security, vol. 16, pp. 1452–1466, 2021.
[21]
X.-C. Li, X.-Y. Zhang, F. Yin, and C.-L. Liu, “Decision-based adversarial attack with frequency mixup,” IEEE Trans. Inf. Forensics Security, vol. 17, pp. 1038–1052, 2022.
[22]
X. Wang, Z. Zhang, G. Lu, and Y. Xu, “Targeted attack and defense for deep hashing,” in Proc. 44th Int. ACM SIGIR Conf. Res. Develop. Inf. Retr., Jul. 2021, pp. 2298–2302.
[23]
A. Madry, A. Makelov, L. Schmidt, D. Tsipras, and A. Vladu, “Towards deep learning models resistant to adversarial attacks,” in Proc. Int. Conf. Learn. Represent., 2017, pp. 1–28.
[24]
X. Luoet al., “A survey on deep hashing methods,” ACM Trans. Knowl. Discovery From Data, vol. 17, no. 1, pp. 1–50, Feb. 2023.
[25]
R. Salakhutdinov and G. Hinton, “Semantic hashing,” Int. J. Approx. Reasoning, vol. 50, no. 7, pp. 969–978, Jul. 2009.
[26]
K. G. Dizaji, F. Zheng, N. S. Nourabadi, Y. Yang, C. Deng, and H. Huang, “Unsupervised deep generative adversarial hashing network,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018, pp. 3664–3673.
[27]
A. W. M. Smeulders, M. Worring, S. Santini, A. Gupta, and R. Jain, “Content-based image retrieval at the end of the early years,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 22, no. 12, pp. 1349–1380, 2000.
[28]
I. J. Goodfellow, J. Shlens, and C. Szegedy, “Explaining and harnessing adversarial examples,” in Proc. Int. Conf. Learn. Represent., 2015, pp. 1–10.
[29]
S.-M. Moosavi-Dezfooli, A. Fawzi, and P. Frossard, “DeepFool: A simple and accurate method to fool deep neural networks,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2016, pp. 2574–2582.
[30]
A. Kurakin, I. J. Goodfellow, and S. Bengio, “Adversarial examples in the physical world,” in Artificial Intelligence Safety and Security. London, U.K.: Chapman & Hall, 2018, pp. 99–112.
[31]
J. Lin, C. Song, K. He, L. Wang, and J. E. Hopcroft, “Nesterov accelerated gradient and scale invariance for adversarial attacks,” 2019, arXiv:1908.06281.
[32]
F. Croce and M. Hein, “Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks,” in Proc. Int. Conf. Mach. Learn., 2020, pp. 2206–2216.
[33]
Q. Xu, G. Tao, and X. Zhang, “Bounded adversarial attack on deep content features,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2022, pp. 15182–15191.
[34]
C. Xieet al., “Improving transferability of adversarial examples with input diversity,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2019, pp. 2725–2734.
[35]
X. Wang and K. He, “Enhancing the transferability of adversarial attacks through variance tuning,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2021, pp. 1924–1933.
[36]
Z. Yuan, J. Zhang, and S. Shan, “Adaptive image transformations for transfer-based adversarial attack,” in Proc. Eur. Conf. Comput. Vis. Tel Aviv, Israel: Springer, Oct. 2022, pp. 1–17.
[37]
C. Xie, J. Wang, Z. Zhang, Z. Ren, and A. Yuille, “Mitigating adversarial effects through randomization,” 2017, arXiv:1711.01991.
[38]
X. Jia, X. Wei, X. Cao, and H. Foroosh, “ComDefend: An efficient image compression model to defend adversarial examples,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2019, pp. 6077–6085.
[39]
H. Zhang, Y. Yu, J. Jiao, E. Xing, L. El Ghaoui, and M. Jordan, “Theoretically principled trade-off between robustness and accuracy,” in Proc. Int. Conf. Mach. Learn., 2019, pp. 7472–7482.
[40]
D. Wu, S.-T. Xia, and Y. Wang, “Adversarial weight perturbation helps robust generalization,” in Proc. Adv. Neural Inf. Process. Syst., vol. 33, 2020, pp. 2958–2969.
[41]
J. Cui, S. Liu, L. Wang, and J. Jia, “Learnable boundary guided adversarial training,” in Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), Oct. 2021, pp. 15701–15710.
[42]
X. Jia, Y. Zhang, B. Wu, K. Ma, J. Wang, and X. Cao, “LAS-AT: Adversarial training with learnable attack strategy,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2022, pp. 13388–13398.
[43]
T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A simple framework for contrastive learning of visual representations,” in Proc. Int. Conf. Mach. Learn., 2020, pp. 1597–1607.
[44]
P. Khoslaet al., “Supervised contrastive learning,” in Proc. Adv. Neural Inf. Process. Syst., vol. 33, 2020, pp. 18661–18673.
[45]
M. J. Huiskes and M. S. Lew, “The MIR Flickr retrieval evaluation,” in Proc. 1st ACM Int. Conf. Multimedia Inf. Retr., Oct. 2008, pp. 39–43.
[46]
T.-S. Chua, J. Tang, R. Hong, H. Li, Z. Luo, and Y. Zheng, “NUS-WIDE: A real-world web image database from National University of Singapore,” in Proc. ACM Int. Conf. Image Video Retr., Jul. 2009, pp. 1–9.
[47]
T. Linet al., “Microsoft COCO: Common objects in context,” in Proc. Eur. Conf. Comput. Vis., 2014, pp. 740–755.
[48]
A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” in Proc. Adv. Neural Inf. Process. Syst., 2012, pp. 1097–1105.
[49]
K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in Proc. Int. Conf. Learn. Represent., 2015, pp. 1–14.
[50]
H. Robbins, “A stochastic approximation method,” Ann. Math. Statist., vol. 22, no. 3, pp. 400–407, 2007.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image IEEE Transactions on Information Forensics and Security
IEEE Transactions on Information Forensics and Security  Volume 18, Issue
2023
4507 pages

Publisher

IEEE Press

Publication History

Published: 01 January 2023

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 20 Jan 2025

Other Metrics

Citations

Cited By

View all

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media