Backdoor Attacks in Machine Learning

The field of machine learning is moving towards a greater understanding of the risks and vulnerabilities associated with backdoor attacks. Recent research has focused on developing more sophisticated and stealthy attack methods, including one-to-N backdoor frameworks, weak triggers, and multi-modal prompt tuning. These advances have significant implications for the security of deep learning systems, particularly in safety-sensitive domains such as autonomous driving and robotics. Noteworthy papers in this area include: One-to-N Backdoor Attack in 3D Point Cloud via Spherical Trigger, which establishes a theoretical foundation for one-to-N backdoor attacks in 3D vision. BackWeak, which proposes a simple and efficient backdoor attack paradigm using weak triggers and fine-tuning. The 'Sure' Trap, which introduces a compliance-only backdoor that can be used to analyze the security risks of large language models.

Sources

One-to-N Backdoor Attack in 3D Point Cloud via Spherical Trigger

BackWeak: Backdooring Knowledge Distillation Simply with Weak Triggers and Fine-tuning

The 'Sure' Trap: Multi-Scale Poisoning Analysis of Stealthy Compliance-Only Backdoors in Fine-Tuned Large Language Models

Backdoor Attacks on Open Vocabulary Object Detectors via Multi-Modal Prompt Tuning

Enhancing All-to-X Backdoor Attacks with Optimized Target Class Mapping

Robust Defense Strategies for Multimodal Contrastive Learning: Efficient Fine-tuning Against Backdoor Attacks

Uncovering and Aligning Anomalous Attention Heads to Defend Against NLP Backdoor Attacks

Dynamic Black-box Backdoor Attacks on IoT Sensory Data

Steganographic Backdoor Attacks in NLP: Ultra-Low Poisoning and Defense Evasion

Detecting Sleeper Agents in Large Language Models via Semantic Drift Analysis

Built with on top of