Membership Inference Attacks ML Frameworks

Tools and implementations for detecting whether specific data points were used in model training, including attack methods, defenses, and privacy analysis frameworks. Does NOT include general privacy-preserving ML, differential privacy libraries, or other data poisoning/adversarial attacks.

There are 45 membership inference attacks frameworks tracked. 2 score above 50 (established tier). The highest-rated is google/scaaml at 68/100 with 193 stars and 122 monthly downloads.

Get all 45 projects as JSON

curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=ml-frameworks&subcategory=membership-inference-attacks&limit=20"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.

# Framework Score Tier
1 google/scaaml

SCAAML: Side Channel Attacks Assisted with Machine Learning

68
Established
2 Koukyosyumei/AIJack

Security and Privacy Risk Simulator for Machine Learning (arXiv:2312.17667)

50
Established
3 pralab/secml

A Python library for Secure and Explainable Machine Learning

47
Emerging
4 AI-SDC/SACRO-ML

Collection of tools and resources for managing the statistical disclosure...

46
Emerging
5 oss-slu/mithridatium

Mithridatium is a research-driven project aimed at detecting backdoors and...

39
Emerging
6 matteonerini/pin-side-channel-attacks

Machine Learning for PIN Side-Channel Attacks Based on Smartphone Motion Sensors

38
Emerging
7 liuyugeng/ML-Doctor

Code for ML Doctor

38
Emerging
8 ArtLabss/open-data-anonymizer

Python Data Anonymization & Masking Library For Data Science Tasks

36
Emerging
9 microsoft/responsible-ai-toolbox-privacy

A library for statistically estimating the privacy of ML pipelines from...

35
Emerging
10 yonsei-sslab/MIA

🔒 Implementation of Shokri et al(2016) "Membership Inference Attacks against...

33
Emerging
11 stratosphereips/awesome-ml-privacy-attacks

An awesome list of papers on privacy attacks against machine learning

32
Emerging
12 zhoumingyi/ModelObfuscator

Code for our paper "Modelobfuscator: Obfuscating Model Information to...

31
Emerging
13 YujiaBao/ls

Learning to Split for Automatic Bias Detection

30
Emerging
14 brian-lou/Training-Data-Extraction-Attack-on-LLMs

This project explores training data extraction attacks on the LLaMa 7B,...

30
Emerging
15 MinChen00/UnlearningLeaks

Official implementation of "When Machine Unlearning Jeopardizes Privacy"...

29
Experimental
16 allensll/Awesome-Crypto-DNN

List of papers on cryptography assisted deep learning privacy computation

28
Experimental
17 mmalekzadeh/honest-but-curious-nets

Honest-but-Curious Nets: Sensitive Attributes of Private Inputs Can Be...

28
Experimental
18 gnekt/Pirates-of-the-RAG

Pirates of the RAG: Adaptively Attacking LLMs to Leak Knowledge Bases

27
Experimental
19 dahmansphi/attackai

Test tool to simulate two types of poisoning attack on AI model

24
Experimental
20 MichaelTJC96/Label_Flipping_Attack

The project aims to evaluate the vulnerability of Federated Learning systems...

23
Experimental
21 dahmansphi/protectai

Test tool to simulate defense from poisoning attack on AI model

23
Experimental
22 najeebjebreel/lira_analysis

Revisiting the LiRA Membership Inference Attack Under Realistic Assumptions

23
Experimental
23 FRANCYZXZ/Federated-Learning-Security-Backdoor-Attacks-Gradient-Inversion-Unlearning

A comprehensive framework simulating integrity (Backdoor) and privacy...

23
Experimental
24 FRANCYZXZ/federated-backdoor-mitigation

A comprehensive framework simulating integrity (Backdoor) and privacy...

23
Experimental
25 ibhushani/amnesia

🧠 Enterprise-grade Machine Unlearning architecture. Surgically erases data...

22
Experimental
26 Abhishek-yadav04/AgisFL

AgisFL is a cutting-edge, production-ready cybersecurity platform that...

22
Experimental
27 karthik7129/FL-IoT-Threat_detection

This is the end to end Federated learning pipeline for Iot threat detection

22
Experimental
28 yangarbiter/dp-dg

What You See is What You Get: Distributional Generalization for Algorithm...

22
Experimental
29 davidemodolo/malicious_finetuning

Proof of concept demonstrating backdoor injection into fine-tuned LLMs using...

22
Experimental
30 davidemodolo/malicious_llm_finetuning

Proof of concept demonstrating backdoor injection into fine-tuned LLMs using...

22
Experimental
31 X1aoyangXu/FORA

Official code of the paper "A Stealthy Wrongdoer: Feature-Oriented...

21
Experimental
32 VissaMoutafis/Membership-Inference-Research

Bachelor's Thesis on Membership Inference Attacks

21
Experimental
33 ege-erdogan/unsplit

Supplementary code for the paper "UnSplit: Data-Oblivious Model Inversion,...

20
Experimental
34 trucndt/ami

Codebase for Active Membership Inference Attack under Local Differential...

17
Experimental
35 Pilladian/ml-attack-framework

Universität des Saarlandes - Privacy Enhancing Technologies 2021 - Semester Project

17
Experimental
36 Jiaqi0602/adversarial-attack-from-leakage

From Gradient Leakage to Adversarial Attacks in Federated Learning

16
Experimental
37 DoktorC/double-strike-host2026

Official repository of the paper "Double Strike: Breaking...

15
Experimental
38 ljvmiranda921/vs-split

A Python library for creating adversarial splits

14
Experimental
39 VirajM723/MachineUnlearning

Machine unlearning using SISA training to efficiently remove data points...

14
Experimental
40 hardware-fab/Hound

Hound: Locating Cryptographic Primitives in Desynchronized Side-Channel...

13
Experimental
41 hardware-fab/DLaTA

A Deep Learning-assisted Template Attack Against Dynamic Frequency Scaling...

13
Experimental
42 dAI-SY-Group/PRECODE

Source code and demonstration for our paper "PRECODE - A Generic Model...

13
Experimental
43 gongzhimin/Copyright-Protection-Studies-in-Deep-Learning

A repository about literature of copyright protection in deep learning.

12
Experimental
44 AmanPriyanshu/The-Unlearning-Protocol

Choose which data to make your model forget (Unlearn!), but watch out -...

11
Experimental
45 Axelboutie/Deep-Learning-for-Side-Channels-Attacks

This repository provides a model of convutionnal neural network or a MLP to...

11
Experimental