Perceptual Carlini-Wagner Attack: A Robust and Imperceptible Adversarial Attack Using LPIPS
Adversarial attacks on deep neural networks (DNNs) present significant challenges by exploiting model vulnerabilities using perturbations that are often imperceptible to human observers. Traditional approaches typically constrain perturbations using p-norms, which do not effectively capture human pe...
Saved in:
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2025-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/11078278/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1839614788120870912 |
---|---|
author | Liming Fan Anis Salwa Mohd Khairuddin Haichuan Liu Khairunnisa Binti Hasikin |
author_facet | Liming Fan Anis Salwa Mohd Khairuddin Haichuan Liu Khairunnisa Binti Hasikin |
author_sort | Liming Fan |
collection | DOAJ |
description | Adversarial attacks on deep neural networks (DNNs) present significant challenges by exploiting model vulnerabilities using perturbations that are often imperceptible to human observers. Traditional approaches typically constrain perturbations using p-norms, which do not effectively capture human perceptual similarity. In this work, we propose the Perceptual Carlini-Wagner (PCW) attack, which integrates the Learned Perceptual Image Patch Similarity (LPIPS) metric into the adversarial optimization process. By replacing p-norm constraints with LPIPS, PCW generates adversarial examples that are both highly effective at inducing misclassification and visually indistinguishable from the original images. We evaluate PCW on the CIFAR-10, CIFAR-100, and ImageNet datasets. On ImageNet, adversarial examples crafted using PCW achieve an LPIPS distance of only 0.0002 from clean images, in contrast to 0.3 LPIPS for those produced by the CW and PGD attacks. In terms of robustness, PCW shows superior performance under common image processing defenses such as JPEG compression and bit-depth reduction, outperforming CW and SSAH and rivaling PGD. Additionally, we test PCW against adversarially trained models from RobustBench and find that it maintains high attack success rates, significantly outperforming CW and PGD in this more challenging setting. Finally, we assess the transferability of PCW across model architectures. While LPIPS contributes to perceptual alignment, it does not significantly improve transferability, with results comparable to those of the original CW attack. |
format | Article |
id | doaj-art-3a96e9bbe5e54493817ddd4fe7d30ef1 |
institution | Matheson Library |
issn | 2169-3536 |
language | English |
publishDate | 2025-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj-art-3a96e9bbe5e54493817ddd4fe7d30ef12025-07-25T23:00:48ZengIEEEIEEE Access2169-35362025-01-011312918512919410.1109/ACCESS.2025.358811311078278Perceptual Carlini-Wagner Attack: A Robust and Imperceptible Adversarial Attack Using LPIPSLiming Fan0https://orcid.org/0009-0009-6917-6382Anis Salwa Mohd Khairuddin1https://orcid.org/0000-0002-9873-4779Haichuan Liu2https://orcid.org/0000-0003-0559-5552Khairunnisa Binti Hasikin3https://orcid.org/0000-0002-0471-3820Department of Electrical Engineering, Faculty of Engineering, Universiti Malaya, Kuala Lumpur, MalaysiaDepartment of Electrical Engineering, Faculty of Engineering, Universiti Malaya, Kuala Lumpur, MalaysiaDepartment of Electrical Engineering, Faculty of Engineering, Universiti Malaya, Kuala Lumpur, MalaysiaDepartment of Biomedical Engineering, Faculty of Engineering, Universiti Malaya, Kuala Lumpur, MalaysiaAdversarial attacks on deep neural networks (DNNs) present significant challenges by exploiting model vulnerabilities using perturbations that are often imperceptible to human observers. Traditional approaches typically constrain perturbations using p-norms, which do not effectively capture human perceptual similarity. In this work, we propose the Perceptual Carlini-Wagner (PCW) attack, which integrates the Learned Perceptual Image Patch Similarity (LPIPS) metric into the adversarial optimization process. By replacing p-norm constraints with LPIPS, PCW generates adversarial examples that are both highly effective at inducing misclassification and visually indistinguishable from the original images. We evaluate PCW on the CIFAR-10, CIFAR-100, and ImageNet datasets. On ImageNet, adversarial examples crafted using PCW achieve an LPIPS distance of only 0.0002 from clean images, in contrast to 0.3 LPIPS for those produced by the CW and PGD attacks. In terms of robustness, PCW shows superior performance under common image processing defenses such as JPEG compression and bit-depth reduction, outperforming CW and SSAH and rivaling PGD. Additionally, we test PCW against adversarially trained models from RobustBench and find that it maintains high attack success rates, significantly outperforming CW and PGD in this more challenging setting. Finally, we assess the transferability of PCW across model architectures. While LPIPS contributes to perceptual alignment, it does not significantly improve transferability, with results comparable to those of the original CW attack.https://ieeexplore.ieee.org/document/11078278/Adversarial attacksperceptual similarityLPIPSdeep neural networksrobustnesstransferability |
spellingShingle | Liming Fan Anis Salwa Mohd Khairuddin Haichuan Liu Khairunnisa Binti Hasikin Perceptual Carlini-Wagner Attack: A Robust and Imperceptible Adversarial Attack Using LPIPS IEEE Access Adversarial attacks perceptual similarity LPIPS deep neural networks robustness transferability |
title | Perceptual Carlini-Wagner Attack: A Robust and Imperceptible Adversarial Attack Using LPIPS |
title_full | Perceptual Carlini-Wagner Attack: A Robust and Imperceptible Adversarial Attack Using LPIPS |
title_fullStr | Perceptual Carlini-Wagner Attack: A Robust and Imperceptible Adversarial Attack Using LPIPS |
title_full_unstemmed | Perceptual Carlini-Wagner Attack: A Robust and Imperceptible Adversarial Attack Using LPIPS |
title_short | Perceptual Carlini-Wagner Attack: A Robust and Imperceptible Adversarial Attack Using LPIPS |
title_sort | perceptual carlini wagner attack a robust and imperceptible adversarial attack using lpips |
topic | Adversarial attacks perceptual similarity LPIPS deep neural networks robustness transferability |
url | https://ieeexplore.ieee.org/document/11078278/ |
work_keys_str_mv | AT limingfan perceptualcarliniwagnerattackarobustandimperceptibleadversarialattackusinglpips AT anissalwamohdkhairuddin perceptualcarliniwagnerattackarobustandimperceptibleadversarialattackusinglpips AT haichuanliu perceptualcarliniwagnerattackarobustandimperceptibleadversarialattackusinglpips AT khairunnisabintihasikin perceptualcarliniwagnerattackarobustandimperceptibleadversarialattackusinglpips |