Skip to main navigation Skip to search Skip to main content

Adversarial Robust Distillation Method Based on Intensity Correlation Regularization Learning

  • Shuobin Lin
  • , Jieyi Cai
  • , Xiaocheng Fang
  • , Zheng Zhang
  • , Guangming Lu
  • , Bingzhi Chen*
  • *Corresponding author for this work
  • South China Normal University
  • Harbin Institute of Technology

Research output: Contribution to journalArticlepeer-review

Abstract

This research introduces an Adversarial Robust Distillation (ARD) method based on Intensity Correlation Regularization Learning (ICRL) to address the limitations of existing ARD approaches, which are hindered by insufficient and unreliable guidance from the teacher network and fixed attack strengths. The proposed method comprises two key modules: multidimensional knowledge distillation and dynamic adjustment of attack intensity. Multidimensional knowledge distillation effectively addresses distributional discrepancies caused by inadequate or unreliable teacher network guidance by incorporating instance-level and class-level knowledge distillation across teacher and student logits, as well as introspective self-distillation within the student network. To enable adaptive updates of attack strength based on the improved robustness of the student network, an efficient intensity dynamic adjustment algorithm is designed to dynamically select and assign the appropriate attack intensities for each instance. Additionally, ICRL applies regularization to the introspective self-distillation dimension from the attack strength perspective, adaptively normalizing the student's introspective loss and mitigating the impact of extremely adversarial perturbation instances. Extensive experimental results on the CIFAR-10 and CIFAR-100 datasets demonstrated that this method functions as a universal plugin for most mainstream ARD frameworks, and significantly enhances the resilience of baseline methods against multistep attacks. In particular, with the current state-of-the-art baseline method named AdaAD, when using a ResNet-18 student network, AdaAD-ICRL achieves improvements of 2.06 and 2.11 percentage points in adversarial robust accuracy against Projected Gradient Descent (PGD)-10 attacks, validating the compatibility and effectiveness of ICRL within existing frameworks.

Original languageEnglish
Pages (from-to)42-50
Number of pages9
JournalJisuanji Gongcheng/Computer Engineering
Volume51
Issue number1
DOIs
StatePublished - 15 Jan 2025
Externally publishedYes

Keywords

  • Adversarial Training (AT)
  • adversarial robustness
  • dynamic intensity adjustment
  • knowledge distillation
  • regularization

Fingerprint

Dive into the research topics of 'Adversarial Robust Distillation Method Based on Intensity Correlation Regularization Learning'. Together they form a unique fingerprint.

Cite this