Backdoor Attacks & Defense

Explore how adversaries can embed hidden behaviors in ML models, learn to detect trojan models using Neural Cleanse and spectral signatures, and master defense techniques like fine-pruning and activation clustering.

6
Lessons
🔒
Security Focus
🕑
Self-Paced
100%
Free

Your Learning Path

Follow these lessons in order, or jump to any topic that interests you.

What You'll Learn

By the end of this course, you'll be able to:

🔎

Identify Backdoor Threats

Recognize how and where backdoors can be inserted into ML models, datasets, and training pipelines.

🛡

Detect Trojan Models

Apply state-of-the-art detection methods including Neural Cleanse, spectral signatures, and activation analysis.

🛠

Remove Backdoors

Use fine-pruning, distillation, and unlearning techniques to neutralize backdoors from compromised models.

🎯

Secure ML Supply Chains

Implement policies and technical controls to prevent backdoor injection throughout the ML lifecycle.