Official PyTorch implementation of "Scanning Trojaned Models Using Out-of-Distribution Samples" (NeurIPS 2024)
-
Updated
Oct 26, 2024 - Jupyter Notebook
Official PyTorch implementation of "Scanning Trojaned Models Using Out-of-Distribution Samples" (NeurIPS 2024)
School AI semester project
This repository provide the studies on the security of language models for code (CodeLMs).
A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
AIShield Watchtower: Dive Deep into AI's Secrets! 🔍 Open-source tool by AIShield for AI model insights & vulnerability scans. Secure your AI supply chain today! ⚙️🛡️
M.Sc. thesis work on adversarial attacks against anti-spoofing models.
TransferAttack is a pytorch framework to boost the adversarial transferability for image classification.
Fantastic Robustness Measures: The Secrets of Robust Generalization [NeurIPS 2023]
🛡 A set of adversarial attacks in PyTorch
privat1 :(privatize) is a toolset to anonimize images and prevent object recognition.
Repository accompanying the paper https://arxiv.org/abs/2407.14937
Code and Data for the paper "LPF-Defense: 3D Adversarial Defense based on Frequency Analysis", PLoS ONE
The official implementation of ECCV'24 paper "To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Unsafe Images ... For Now". This work introduces one fast and effective attack method to evaluate the harmful-content generation ability of safety-driven unlearned diffusion models.
Adversary Emulation Framework
This is a code repository for a paper with title "Adversarial Attack and Defence of Federated Learning-Based Network Traffic Classification in Edge Computing Environment"
A graph reliability toolbox based on PyTorch and PyTorch Geometric (PyG).
DeepDefend is an open-source Python library for adversarial attacks and defenses in deep learning models, enhancing the security and robustness of AI systems.
Official repo for the code implementation of the paper "DiffusionGuard: A Robust Defense Against Malicious Diffusion-based Image Editing"
"Neural Computing and Applications" Published Paper (2023)
Add a description, image, and links to the adversarial-attacks topic page so that developers can more easily learn about it.
To associate your repository with the adversarial-attacks topic, visit your repo's landing page and select "manage topics."