Attack as Detection: Using Adversarial Attack Methods to Detect Abnormal Examples
2023-11
发表期刊ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY (IF:6.6[JCR-2023],6.6[5-Year])
ISSN1049-331X
EISSN1557-7392
卷号33期号:3
发表状态已发表
DOI10.1145/3631977
摘要

As a new programming paradigm, deep learning (DL) has achieved impressive performance in areas such as image processing and speech recognition, and has expanded its application to solve many real-world problems. However, neural networks and DL are normally black-box systems; even worse, DL-based software are vulnerable to threats from abnormal examples, such as adversarial and backdoored examples constructed by attackers with malicious intentions as well as unintentionally mislabeled samples. Therefore, it is important and urgent to detect such abnormal examples. Although various detection approaches have been proposed respectively addressing some specific types of abnormal examples, they suffer from some limitations; until today, this problem is still of considerable interest. In this work, we first propose a novel characterization to distinguish abnormal examples from normal ones based on the observation that abnormal examples have significantly different (adversarial) robustness from normal ones. We systemically analyze those three different types of abnormal samples in terms of robustness and find that they have different characteristics from normal ones. As robustness measurement is computationally expensive and hence can be challenging to scale to large networks, we then propose to effectively and efficiently measure robustness of an input sample using the cost of adversarially attacking the input, which was originally proposed to test robustness of neural networks against adversarial examples. Next, we propose a novel detection method, named attack as detection (A2D for short), which uses the cost of adversarially attacking an input instead of robustness to check if it is abnormal. Our detection method is generic, and various adversarial attack methods could be leveraged. Extensive experiments show that A2D is more effective than recent promising approaches that were proposed to detect only one specific type of abnormal examples. We also thoroughly discuss possible adaptive attack methods to our adversarial example detection method and show that A2D is still effective in defending carefully designed adaptive adversarial attack methods - for example, the attack success rate drops to 0% on CIFAR10. © 2024 Copyright held by the owner/author(s). Publication rights licensed to ACM.

关键词Deep learning Image processing Adversarial example Attack methods Backdoored sample Deep learning Detection Detection methods Mislabeled sample Neural-networks Performance Programming paradigms
收录类别SCI ; EI
语种英语
出版者Association for Computing Machinery
EI入藏号20241816023862
EI主题词Speech recognition
EI分类号461.4 Ergonomics and Human Factors Engineering ; 723.2 Data Processing and Image Processing ; 751.5 Speech
原始文献类型Journal article (JA)
引用统计
正在获取...
文献类型期刊论文
条目标识符https://kms.shanghaitech.edu.cn/handle/2MSLDSTB/346052
专题信息科学与技术学院_PI研究组_宋富组
信息科学与技术学院_本科生
信息科学与技术学院_博士生
通讯作者Song, Fu
作者单位
1.ShanghaiTech University
2.State Key Laboratory of Computer Science, Institute of Software, Chinese Academy of Sciences
3.University of Chinese Academy of Sciences
4.Zhejiang University
5.Singapore Management University, Singapore
第一作者单位上海科技大学
第一作者的第一单位上海科技大学
推荐引用方式
GB/T 7714
Zhao, Zhe,Chen, Guangke,Liu, Tong,et al. Attack as Detection: Using Adversarial Attack Methods to Detect Abnormal Examples[J]. ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY,2023,33(3).
APA Zhao, Zhe.,Chen, Guangke.,Liu, Tong.,Li, Taishan.,Song, Fu.,...&Sun, Jun.(2023).Attack as Detection: Using Adversarial Attack Methods to Detect Abnormal Examples.ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY,33(3).
MLA Zhao, Zhe,et al."Attack as Detection: Using Adversarial Attack Methods to Detect Abnormal Examples".ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY 33.3(2023).
条目包含的文件
文件名称/大小 文献类型 版本类型 开放类型 使用许可
个性服务
查看访问统计
谷歌学术
谷歌学术中相似的文章
[Zhao, Zhe]的文章
[Chen, Guangke]的文章
[Liu, Tong]的文章
百度学术
百度学术中相似的文章
[Zhao, Zhe]的文章
[Chen, Guangke]的文章
[Liu, Tong]的文章
必应学术
必应学术中相似的文章
[Zhao, Zhe]的文章
[Chen, Guangke]的文章
[Liu, Tong]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。