ArXiv TLDR

Defending Quantum Classifiers against Adversarial Perturbations through Quantum Autoencoders

🐦 Tweet
2604.28176

Emma Andrews, Sahan Sanjaya, Prabhat Mishra

quant-phcs.LG

TLDR

This paper introduces a quantum autoencoder-based defense that purifies adversarial samples to protect quantum classifiers without adversarial training.

Key contributions

  • Introduces an adversarial training-free defense framework for quantum classifiers.
  • Employs a quantum autoencoder to purify adversarial samples through reconstruction.
  • Offers a confidence metric to identify adversarial samples that cannot be purified.
  • Demonstrates up to 68% prediction accuracy improvement against adversarial attacks.

Why it matters

Quantum machine learning models are vulnerable to adversarial attacks, and current defenses have practical limitations. This paper offers a novel, training-free approach using quantum autoencoders to secure these models. It significantly improves robustness, making quantum classifiers more reliable for real-world applications.

Original Abstract

Machine learning models can learn from data samples to carry out various tasks efficiently. When data samples are adversarially manipulated, such as by insertion of carefully crafted noise, it can cause the model to make mistakes. Quantum machine learning models are also vulnerable to such adversarial attacks, especially in image classification using variational quantum classifiers. While there are promising defenses against these adversarial perturbations, such as training with adversarial samples, they face practical limitations. For example, they are not applicable in scenarios where training with adversarial samples is either not possible or can overfit the models on one type of attack. In this paper, we propose an adversarial training-free defense framework that utilizes a quantum autoencoder to purify the adversarial samples through reconstruction. Moreover, our defense framework provides a confidence metric to identify potentially adversarial samples that cannot be purified the quantum autoencoder. Extensive evaluation demonstrates that our defense framework can significantly outperform state-of-the-art in prediction accuracy (up to 68%) under adversarial attacks.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.