[go: up one dir, main page]

Skip to main content

Invisible Backdoor Attacks on Key Regions Based on Target Neurons in Self-Supervised Learning

  • Conference paper
  • First Online:
Knowledge Science, Engineering and Management (KSEM 2024)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 14886))

  • 434 Accesses

Abstract

With the emergence and popularity of large-scale unlabeled data, self-supervised learning has become a new development trend. This learning paradigm can acquire high-quality representations of complex data without relying on labels, saving labor costs, avoiding human labeling errors, and rivaling supervised learning in expressing image features. However, research has revealed insufficient security in neural network models trained under the self-supervised learning paradigm, making them susceptible to backdoor attacks. This malicious behavior significantly undermines the security of artificial intelligence models and severely hinders the development of the Internet of Things with intelligence. This paper proposes an invisible backdoor attack scheme on key regions based on target neurons. Firstly, the key regions of image feature expression are determined through the attention mechanism. In this region, this paper trains a set of critical neurons to obtain a trigger capable of inducing misclassifications. Subsequently, a poison dataset is constructed to attack the self-supervised training model. The triggers generated by this scheme resemble random noise. They are inconspicuous in visual space, achieving a high attack success rate and enhancing both the concealment and effectiveness of the triggers. The self-supervised training model with the implanted backdoor evades backdoor detection, further increasing the model’s indistinguishability. Ultimately, experimental results demonstrate that while ensuring the concealment of triggers, this scheme can achieve a high attack success rate with only 1% poison data, and the poison model can escape detection.

X. Qian and Y. He—Contribute equally to this work.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 199.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Chen, B., et al.: Detecting backdoor attacks on deep neural networks by activation clustering. arXiv preprint arXiv:1811.03728 (2018)

  2. Chen, K., Hong, L., Xu, H., Li, Z., Yeung, D.Y.: Multisiam: self-supervised multi-instance siamese representation learning for autonomous driving. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 7546–7554 (2021)

    Google Scholar 

  3. Chen, X., He, K.: Exploring simple siamese representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15750–15758 (2021)

    Google Scholar 

  4. Chen, Y., et al.: Image super-resolution reconstruction based on feature map attention mechanism. Appl. Intell. 51, 4367–4380 (2021)

    Article  Google Scholar 

  5. Chou, S.Y., Chen, P.Y., Ho, T.Y.: How to backdoor diffusion models? In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4015–4024 (2023)

    Google Scholar 

  6. Doan, K., Lao, Y., Zhao, W., Li, P.: Lira: Learnable, imperceptible and robust backdoor attacks. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 11966–11976 (2021)

    Google Scholar 

  7. Gu, T., Liu, K., Dolan-Gavitt, B., Garg, S.: Badnets: evaluating backdooring attacks on deep neural networks. IEEE Access 7, 47230–47244 (2019)

    Article  Google Scholar 

  8. Gulzar, Y.: Fruit image classification model based on mobilenetv2 with deep transfer learning technique. Sustainability 15, 1906 (2023)

    Article  Google Scholar 

  9. Lee, J.H., Kim, H., Park, H.J., Heo, J.H.: Temporal prediction modeling for rainfall-induced shallow landslide hazards using extreme value distribution. Landslides 18, 321–338 (2021)

    Article  Google Scholar 

  10. Li, Y., Lyu, X., Koren, N., Lyu, L., Li, B., Ma, X.: Anti-backdoor learning: training clean models on poisoned data. Adv. Neural. Inf. Process. Syst. 34, 14900–14912 (2021)

    Google Scholar 

  11. Li, Y., Li, Y., Wu, B., Li, L., He, R., Lyu, S.: Invisible backdoor attack with sample-specific triggers. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 16463–16472 (2021)

    Google Scholar 

  12. Liang, H., et al.: Self-supervised spatiotemporal representation learning by exploiting video continuity. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 1564–1573 (2022)

    Google Scholar 

  13. Misra, I., Maaten, L.v.d.: Self-supervised learning of pretext-invariant representations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6707–6717 (2020)

    Google Scholar 

  14. Niizumi, D., Takeuchi, D., Ohishi, Y., Harada, N., Kashino, K.: Byol for audio: self-supervised learning for general-purpose audio representation. In: 2021 International Joint Conference on Neural Networks (IJCNN), pp. 1–8 (2021)

    Google Scholar 

  15. Saha, A., Tejankar, A., Koohpayegani, S.A., Pirsiavash, H.: Backdoor attacks on self-supervised learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13337–13346 (2022)

    Google Scholar 

  16. Wang, B., et al.: Neural cleanse: Identifying and mitigating backdoor attacks in neural networks. In: 2019 IEEE Symposium on Security and Privacy (SP), pp. 707–723 (2019)

    Google Scholar 

  17. Wang, Y., Liu, Y., Wang, Q., Wang, C., Li, C.: Poisoning self-supervised learning based sequential recommendations. In: Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 300–310 (2023)

    Google Scholar 

  18. Xu, J., et al..: Groupvit: semantic segmentation emerges from text supervision. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 18134–18144 (2022)

    Google Scholar 

  19. Zeng, Y., Park, W., Mao, Z.M., Jia, R.: Rethinking the backdoor attacks’ triggers: a frequency perspective. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 16473–16481 (2021)

    Google Scholar 

  20. Zhao, S., Ma, X., Zheng, X., Bailey, J., Chen, J., Jiang, Y.G.: Clean-label backdoor attacks on video recognition models. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14443–14452 (2020)

    Google Scholar 

  21. Zhao, Y., Wang, G., Luo, C., Zeng, W., Zha, Z.J.: Self-supervised visual representations learning by contrastive mask prediction. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 10160–10169 (2021)

    Google Scholar 

Download references

Acknowledgments

This research is supported by the National Natural Science Foundation of China (NSFC) under grant number 62172377, the Taishan Scholars Program of Shandong province under grant number tsqn202312102, and the Startup Research Foundation for Distinguished Scholars under grant number 202112016.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hui Xia .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Qian, X., He, Y., Zhang, R., Kang, Z., Sheng, Y., Xia, H. (2024). Invisible Backdoor Attacks on Key Regions Based on Target Neurons in Self-Supervised Learning. In: Cao, C., Chen, H., Zhao, L., Arshad, J., Asyhari, T., Wang, Y. (eds) Knowledge Science, Engineering and Management. KSEM 2024. Lecture Notes in Computer Science(), vol 14886. Springer, Singapore. https://doi.org/10.1007/978-981-97-5498-4_10

Download citation

  • DOI: https://doi.org/10.1007/978-981-97-5498-4_10

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-97-5497-7

  • Online ISBN: 978-981-97-5498-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics