Publication | Closed Access
Defending and Harnessing the Bit-Flip Based Adversarial Weight Attack
77
Citations
21
References
2020
Year
Unknown Venue
Convolutional Neural NetworkEngineeringMachine LearningAdversarial AttackEvasion TechniqueInformation SecurityInformation ForensicsBit-flip AttackFault-tolerance CapabilityAdversarial Machine LearningComputer EngineeringComputer ScienceDeep LearningModel CompressionData SecurityCryptographyAdversarial Weight AttackGenerative Adversarial NetworkAttack Model
Recently, a new paradigm of the adversarial attack on the quantized neural network weights has attracted great attention, namely, the Bit-Flip based adversarial weight attack, aka. Bit-Flip Attack (BFA). BFA has shown extraordinary attacking ability, where the adversary can malfunction a quantized Deep Neural Network (DNN) as a random guess, through malicious bit-flips on a small set of vulnerable weight bits (e.g., 13 out of 93 millions bits of 8-bit quantized ResNet-18). However, there are no effective defensive methods to enhance the fault-tolerance capability of DNN against such BFA. In this work, we conduct comprehensive investigations on BFA and propose to leverage binarization-aware training and its relaxation - piece-wise clustering as simple and effective countermeasures to BFA. The experiments show that, for BFA to achieve the identical prediction accuracy degradation (e.g., below 11% on CIFAR-10), it requires 19.3× and 480.1× more effective malicious bit-flips on ResNet-20 and VGG-11 respectively, compared to defend-free counterparts.
| Year | Citations | |
|---|---|---|
Page 1
Page 1