计算机与现代化 ›› 2022, Vol. 0 ›› Issue (11): 52-59.

• 信息安全 • 上一篇    下一篇

基于图像着色的无限制攻击

  

  1. (1.中国石油大学(华东)海洋与空间信息学院,山东青岛266580;2.中国石油大学(华东)计算机科学与技术学院,山东青岛266580)
  • 出版日期:2022-11-30 发布日期:2022-11-30
  • 作者简介:李世宝(1978—),男,山东潍坊人,副教授,硕士生导师,硕士,研究方向:移动计算,干扰对齐,E-mail: lishibao@upc.edu.cn; 王杰伟(1997—),男,山东济宁人,硕士研究生,研究方向:对抗攻击,E-mail: 419582978@qq.com; 崔学荣(1979—),男,教授,硕士生导师,博士,研究方向:智能感知,E-mail: cuixuerong@163.com。
  • 基金资助:
    国家自然科学基金资助项目(61972417, 61872385, 61902431, 91938204); 中央高校基本科研业务费专项资金资助项目(19CX05003A-4)

Unrestricted Attack Based on Colorization

  1. (1. College of Oceanography and Space Informatics, China University of Petroleum (East China), Qingdao 266580, China;
    2. College of Computer Science and Technology, China University of Petroleum (East China), Qingdao 266580, China)
  • Online:2022-11-30 Published:2022-11-30

摘要: 深度学习目前被广泛应用于计算机视觉、机器人技术和自然语言处理等领域。然而,已有研究表明,深度神经网络在对抗样本面前很脆弱,一个精心制作的对抗样本就可以使深度学习模型判断出错。现有的研究大多通过产生微小的Lp范数扰动来误导分类器的对抗性攻击,但是取得的效果并不理想。本文提出一种新的对抗攻击方法——图像着色攻击,将输入样本转为灰度图,设计一种灰度图上色方法指导灰度图着色,最终利用经过上色的图像欺骗分类器实现无限制攻击。实验表明,这种方法制作的对抗样本在欺骗几种最先进的深度神经网络图像分类器方面有不俗表现,并且通过了人类感知研究测试。

关键词: 对抗攻击, 灰度图着色, 对抗样本, 无限制攻击

Abstract: Deep learning is now widely used in areas such as computer vision, robotics, and natural language processing. However, it has been shown that deep neural networks are vulnerable to adversarial examples, and a single carefully crafted adversarial example can make deep learning models misjudge. Most of the existing studies mislead the adversarial attack on classifiers by generating a small perturbation of the Lp paradigm, but the results achieved are not satisfactory. In this paper, we propose a new adversarial attack method, colorization adversarial attack, which converts the input samples into grayscale maps, designs a grayscale coloring method to guide the grayscale map coloring, and finally uses the colorized images to deceive the classifier to achieve unrestricted attacks. Experiments show that the adversarial examples produced by this method performs well in deceiving several state-of-the-art deep neural network image classifiers and passes human perception research tests.

Key words: adversarial attack, colorization, adversarial examples, unrestricted attack