ASSET
This repository is the official implementation of the paper "ASSET: Robust Backdoor Data Detection Across a Multiplicity of Deep Learning Paradigms." ASSET achieves state-of-the-art reliability in detecting poisoned samples in end-to-end supervised learning/ self-supervised learning/ transfer learning.
Install / Use
/learn @reds-lab/ASSETREADME
ASSET: Robust Backdoor Data Detection Across a Multiplicity of Deep Learning Paradigms
This repository is the official implementation of the Usenix Security 2023 paper "ASSET: Robust Backdoor Data Detection Across a Multiplicity of Deep Learning Paradigms." We find that existing detection methods cannot be applied or suffer limited performance for Self-Supervised Learning and transfer learning; even for the widely studied end-to-end supervised learning setting, there is still large room to improve detection in terms of their robustness to variations in poison ratio and attack designs.
To address this problem...actively introduce diffrent model behaviors...
Features
<div align=center> <img src="https://user-images.githubusercontent.com/77789132/218583421-1184b200-5dd0-418a-82a7-15754704fc2f.png"> </div>In the past, the detection of backdoor data was primarily researched within the framework of end-to-end supervised learning (SL). However, in recent years, the use of self-supervised learning (SSL) and transfer learning (TL) has become increasingly popular due to their reduced requirement for labeled data. It has also been shown that successful backdoor attacks can be carried out in these novel settings. Wepropose a new detection method called Active Separation via Offset (ASSET), which actively induces different model behaviors between the backdoor and clean samples to promote their separation. ASSET enables stable defense under different learning paradigms.

Requirements
- Python >= 3.6
- PyTorch >= 1.10.1
- Torchvision >= 0.11.2
- Imageio >= 2.9.0
Usage & HOW-TO
<p align="justify">Use the ASSET_demo.ipynb notebook for a quick start of the ASSET defense (demonstrated on the CIFAR-10 dataset). The default setting running on the CIFAR-10 dataset and attack method is BadNets on ResNet-18.</p>Can you make it easier?
Related Skills
best-practices-researcher
The most comprehensive Claude Code skills registry | Web Search: https://skills-registry-web.vercel.app
groundhog
398Groundhog's primary purpose is to teach people how Cursor and all these other coding agents work under the hood. If you understand how these coding assistants work from first principles, then you can drive these tools harder (or perhaps make your own!).
isf-agent
a repo for an agent that helps researchers apply for isf funding
last30days-skill
17.2kAI agent skill that researches any topic across Reddit, X, YouTube, HN, Polymarket, and the web - then synthesizes a grounded summary
