Evading deepfake detectors via adversarial statistical consistency

In recent years, as various realistic face forgery techniques known as DeepFake improves by leaps and bounds, more and more DeepFake detection techniques have been proposed. These methods typically rely on detecting statistical differences between natural (i.e., real) and DeepFake-generated images i...

Full description

Saved in:
Bibliographic Details
Main Authors: HOU, Yang, GUO, Qing, HUANG, Yihao, XIE, Xiaofei, MA, Lei, ZHAO, Jianjun
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8230
https://ink.library.smu.edu.sg/context/sis_research/article/9233/viewcontent/evading_deepfake_detectors_va_adversarial_statistical_consistency.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9233
record_format dspace
spelling sg-smu-ink.sis_research-92332023-10-26T03:44:22Z Evading deepfake detectors via adversarial statistical consistency HOU, Yang GUO, Qing HUANG, Yihao XIE, Xiaofei MA, Lei ZHAO, Jianjun In recent years, as various realistic face forgery techniques known as DeepFake improves by leaps and bounds, more and more DeepFake detection techniques have been proposed. These methods typically rely on detecting statistical differences between natural (i.e., real) and DeepFake-generated images in both spatial and frequency domains. In this work, we propose to explicitly minimize the statistical differences to evade state-of-the-art DeepFake detectors. To this end, we propose a statistical consistency attack (StatAttack) against DeepFake detectors, which contains two main parts. First, we select several statistical-sensitive natural degradations (i.e., exposure, blur, and noise) and add them to the fake images in an adversarial way. Second, we find that the statistical differences between natural and DeepFake images are positively associated with the distribution shifting between the two kinds of images, and we propose to use a distribution-aware loss to guide the optimization of different degradations. As a result, the feature distributions of generated adversarial examples is close to the natural images. Furthermore, we extend the StatAttack to a more powerful version, MStatAttack, where we extend the single-layer degradation to multi-layer degradations sequentially and use the loss to tune the combination weights jointly. Comprehensive experimental results on four spatial-based detectors and two frequency-based detectors with four datasets demonstrate the effectiveness of our proposed attack method in both white-box and black-box settings. 2023-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8230 info:doi/10.1109/CVPR52729.2023.01181 https://ink.library.smu.edu.sg/context/sis_research/article/9233/viewcontent/evading_deepfake_detectors_va_adversarial_statistical_consistency.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University degradation deepfakes frequency-domain analysis closed box detectors forgery pattern recognition Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic degradation
deepfakes
frequency-domain analysis
closed box
detectors
forgery
pattern recognition
Graphics and Human Computer Interfaces
spellingShingle degradation
deepfakes
frequency-domain analysis
closed box
detectors
forgery
pattern recognition
Graphics and Human Computer Interfaces
HOU, Yang
GUO, Qing
HUANG, Yihao
XIE, Xiaofei
MA, Lei
ZHAO, Jianjun
Evading deepfake detectors via adversarial statistical consistency
description In recent years, as various realistic face forgery techniques known as DeepFake improves by leaps and bounds, more and more DeepFake detection techniques have been proposed. These methods typically rely on detecting statistical differences between natural (i.e., real) and DeepFake-generated images in both spatial and frequency domains. In this work, we propose to explicitly minimize the statistical differences to evade state-of-the-art DeepFake detectors. To this end, we propose a statistical consistency attack (StatAttack) against DeepFake detectors, which contains two main parts. First, we select several statistical-sensitive natural degradations (i.e., exposure, blur, and noise) and add them to the fake images in an adversarial way. Second, we find that the statistical differences between natural and DeepFake images are positively associated with the distribution shifting between the two kinds of images, and we propose to use a distribution-aware loss to guide the optimization of different degradations. As a result, the feature distributions of generated adversarial examples is close to the natural images. Furthermore, we extend the StatAttack to a more powerful version, MStatAttack, where we extend the single-layer degradation to multi-layer degradations sequentially and use the loss to tune the combination weights jointly. Comprehensive experimental results on four spatial-based detectors and two frequency-based detectors with four datasets demonstrate the effectiveness of our proposed attack method in both white-box and black-box settings.
format text
author HOU, Yang
GUO, Qing
HUANG, Yihao
XIE, Xiaofei
MA, Lei
ZHAO, Jianjun
author_facet HOU, Yang
GUO, Qing
HUANG, Yihao
XIE, Xiaofei
MA, Lei
ZHAO, Jianjun
author_sort HOU, Yang
title Evading deepfake detectors via adversarial statistical consistency
title_short Evading deepfake detectors via adversarial statistical consistency
title_full Evading deepfake detectors via adversarial statistical consistency
title_fullStr Evading deepfake detectors via adversarial statistical consistency
title_full_unstemmed Evading deepfake detectors via adversarial statistical consistency
title_sort evading deepfake detectors via adversarial statistical consistency
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8230
https://ink.library.smu.edu.sg/context/sis_research/article/9233/viewcontent/evading_deepfake_detectors_va_adversarial_statistical_consistency.pdf
_version_ 1781793968205332480