Introduction

This is the homepage of our paper De-AntiFake: Rethinking the Protective Perturbations Against Voice Cloning Attacks. This work presents a systematic evaluation of protective perturbations against voice cloning attacks and proposes a novel two-stage purification method to disrupt existing defenses.

Abstract

The rapid advancement of speech generation models has heightened privacy and security concerns related to voice cloning (VC). Recent studies have investigated disrupting unauthorized voice cloning by introducing adversarial perturbations. However, determined attackers can mitigate these protective perturbations and successfully execute VC.

Overview of voice cloning protection and attack scenarios

Fig. 1: The threat model: Bypassing voice cloning defenses using purification methods.

In this study, we conduct the first systematic evaluation of these protective perturbations against VC under realistic threat models that include perturbation purification. Our findings reveal that while existing purification methods can neutralize a considerable portion of the protective perturbations, they still lead to distortions in the feature space of VC models, which degrades the performance of VC.

PhonePuRe methodology diagram

Fig. 2: Our two-stage method: A refinement stage repairs distortions left by initial purification.

From this perspective, we propose a novel two-stage purification method: (1) Purify the perturbed speech; (2) Refine it using phoneme guidance to align it with the clean speech distribution. Experimental results demonstrate that our method outperforms state-of-the-art purification methods in disrupting VC defenses. Our study reveals the limitations of adversarial perturbation-based VC defenses and underscores the urgent need for more robust solutions to mitigate the security and privacy risks posed by VC.

Experimental results comparison

Fig. 3: Our method consistently achieves the highest attack success rate across all tested defenses,
boosting performance from 45.1% to 76.2% in the most challenging case (AntiFake).

Citation

If you find this work useful, please consider citing our paper:

@inproceedings{de-antifake-icml2025,
  title = {De-AntiFake: Rethinking the Protective Perturbations Against Voice Cloning Attacks},
  author = {Fan, Wei and Chen, Kejiang and Liu, Chang and Zhang, Weiming and Yu, Nenghai},
  booktitle = {International Conference on Machine Learning},
  year = {2025},
}

Contact

Technical Questions: range@mail.ustc.edu.cn.

General Inquiries: chenkj@ustc.edu.cn.