Prompt2Perturb (P2P): Text-Guided Diffusion-Based Adversarial Attacks on Breast Ultrasound Images
Abstract
Deep neural networks (DNNs) offer significant promise for improving breast cancer diagnosis in medical imaging. However, these models are highly susceptible to adversarial attacks--small, imperceptible changes that can mislead classifiers--raising critical concerns about their reliability and security. Traditional attacks rely on fixed-norm perturbations, misaligning with human perception. In contrast, diffusion-based attacks require pre-trained models, demanding substantial data when these models are unavailable, limiting practical use in data-scarce scenarios. In medical imaging, however, this is often unfeasible due to the limited availability of datasets. Building on recent advancements in learnable prompts, we propose Prompt2Perturb (P2P), a novel language-guided attack method capable of generating meaningful attack examples driven by text instructions. During the prompt learning phase, our approach leverages learnable prompts within the text encoder to create subtle, yet impactful, perturbations that remain imperceptible while guiding the model towards targeted outcomes. In contrast to current prompt learning-based approaches, our P2P stands out by directly updating text embeddings, avoiding the need for retraining diffusion models. Further, we leverage the finding that optimizing only the early reverse diffusion steps boosts efficiency while ensuring that the generated adversarial examples incorporate subtle noise, thus preserving ultrasound image quality without introducing noticeable artifacts. We show that our method outperforms state-of-the-art attack techniques across three breast ultrasound datasets in FID and LPIPS. Moreover, the generated images are both more natural in appearance and more effective compared to existing adversarial attacks. Our code will be publicly available https://github.com/yasamin-med/P2P.
Community
The paper proposes Prompt2Perturb (P2P), a novel language-guided adversarial attack method that generates imperceptible yet effective perturbations in medical images using text-based prompts, without requiring retraining of diffusion models. By optimizing only the early reverse diffusion steps, P2P ensures high-quality adversarial examples that outperform existing techniques while preserving image fidelity in breast ultrasound datasets.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Boosting the Targeted Transferability of Adversarial Examples via Salient Region&Weighted Feature Drop (2024)
- Visual Adversarial Attack on Vision-Language Models for Autonomous Driving (2024)
- Safeguarding Text-to-Image Generation via Inference-Time Prompt-Noise Optimization (2024)
- Adversarial Prompt Distillation for Vision-Language Models (2024)
- AdvI2I: Adversarial Image Attack on Image-to-Image Diffusion models (2024)
- TAPT: Test-Time Adversarial Prompt Tuning for Robust Inference in Vision-Language Models (2024)
- Adversarial Attacks of Vision Tasks in the Past 10 Years: A Survey (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper