Projects in Awesome Lists tagged with adversarial-attacks
A curated list of projects in awesome lists tagged with adversarial-attacks .
https://github.com/bishopfox/sliver
Adversary Emulation Framework
adversarial-attacks adversary-simulation c2 command-and-control dns dns-server golang gplv3 http implant red-team red-team-engagement red-teaming security-tools sliver
Last synced: 13 Apr 2025
https://github.com/BishopFox/sliver
Adversary Emulation Framework
adversarial-attacks adversary-simulation c2 command-and-control dns dns-server golang gplv3 http implant red-team red-team-engagement red-teaming security-tools sliver
Last synced: 15 Mar 2025
https://github.com/trusted-ai/adversarial-robustness-toolbox
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
adversarial-attacks adversarial-examples adversarial-machine-learning ai artificial-intelligence attack blue-team evasion extraction inference machine-learning poisoning privacy python red-team trusted-ai trustworthy-ai
Last synced: 08 Apr 2025
https://github.com/Trusted-AI/adversarial-robustness-toolbox
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
adversarial-attacks adversarial-examples adversarial-machine-learning ai artificial-intelligence attack blue-team evasion extraction inference machine-learning poisoning privacy python red-team trusted-ai trustworthy-ai
Last synced: 23 Mar 2025
https://github.com/ibm/adversarial-robustness-toolbox
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
adversarial-attacks adversarial-examples adversarial-machine-learning ai artificial-intelligence attack blue-team evasion extraction inference machine-learning poisoning privacy python red-team trusted-ai trustworthy-ai
Last synced: 31 Dec 2024
https://github.com/makcedward/nlpaug
Data augmentation for NLP
adversarial-attacks adversarial-example ai artificial-intelligence augmentation data-science machine-learning ml natural-language-processing nlp
Last synced: 08 Apr 2025
https://github.com/qdata/textattack
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
adversarial-attacks adversarial-examples adversarial-machine-learning data-augmentation machine-learning natural-language-processing nlp security
Last synced: 09 Apr 2025
https://github.com/QData/TextAttack
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
adversarial-attacks adversarial-examples adversarial-machine-learning data-augmentation machine-learning natural-language-processing nlp security
Last synced: 02 Apr 2025
https://github.com/bethgelab/foolbox
A Python toolbox to create adversarial examples that fool neural networks in PyTorch, TensorFlow, and JAX
adversarial-attacks adversarial-examples jax keras machine-learning python pytorch tensorflow
Last synced: 10 Apr 2025
https://github.com/microsoft/promptbench
A unified evaluation framework for large language models
adversarial-attacks benchmark chatgpt evaluation large-language-models prompt prompt-engineering robustness
Last synced: 09 Apr 2025
https://github.com/harry24k/adversarial-attacks-pytorch
PyTorch implementation of adversarial attacks [torchattacks]
adversarial-attacks deep-learning pytorch
Last synced: 11 Apr 2025
https://github.com/Harry24k/adversarial-attacks-pytorch
PyTorch implementation of adversarial attacks [torchattacks]
adversarial-attacks deep-learning pytorch
Last synced: 27 Mar 2025
https://github.com/thunlp/taadpapers
Must-read Papers on Textual Adversarial Attack and Defense
adversarial-attacks adversarial-defense adversarial-learning natural-language-processing nlp paper-list
Last synced: 12 Apr 2025
https://github.com/thunlp/TAADpapers
Must-read Papers on Textual Adversarial Attack and Defense
adversarial-attacks adversarial-defense adversarial-learning natural-language-processing nlp paper-list
Last synced: 27 Mar 2025
https://github.com/baidu/AdvBox
Advbox is a toolbox to generate adversarial examples that fool neural networks in PaddlePaddle、PyTorch、Caffe2、MxNet、Keras、TensorFlow and Advbox can benchmark the robustness of machine learning models. Advbox give a command line tool to generate adversarial examples with Zero-Coding.
adversarial-attacks adversarial-example adversarial-examples deep-learning deepfool fgsm graphpipe machine-learning onnx paddlepaddle security
Last synced: 19 Feb 2025
https://github.com/advboxes/advbox
Advbox is a toolbox to generate adversarial examples that fool neural networks in PaddlePaddle、PyTorch、Caffe2、MxNet、Keras、TensorFlow and Advbox can benchmark the robustness of machine learning models. Advbox give a command line tool to generate adversarial examples with Zero-Coding.
adversarial-attacks adversarial-example adversarial-examples deep-learning deepfool fgsm graphpipe machine-learning onnx paddlepaddle security
Last synced: 08 Apr 2025
https://github.com/advboxes/AdvBox
Advbox is a toolbox to generate adversarial examples that fool neural networks in PaddlePaddle、PyTorch、Caffe2、MxNet、Keras、TensorFlow and Advbox can benchmark the robustness of machine learning models. Advbox give a command line tool to generate adversarial examples with Zero-Coding.
adversarial-attacks adversarial-example adversarial-examples deep-learning deepfool fgsm graphpipe machine-learning onnx paddlepaddle security
Last synced: 15 Mar 2025
https://github.com/borealisai/advertorch
A Toolbox for Adversarial Robustness Research
adversarial-attacks adversarial-example adversarial-examples adversarial-learning adversarial-machine-learning adversarial-perturbations benchmarking machine-learning pytorch robustness security toolbox
Last synced: 14 Apr 2025
https://github.com/BorealisAI/advertorch
A Toolbox for Adversarial Robustness Research
adversarial-attacks adversarial-example adversarial-examples adversarial-learning adversarial-machine-learning adversarial-perturbations benchmarking machine-learning pytorch robustness security toolbox
Last synced: 27 Mar 2025
https://github.com/DSE-MSU/DeepRobust
A pytorch adversarial library for attack and defense methods on images and graphs
adversarial-attacks adversarial-examples deep-learning deep-neural-networks defense graph-convolutional-networks graph-mining graph-neural-networks machine-learning
Last synced: 18 Nov 2024
https://github.com/shubhomoydas/ad_examples
A collection of anomaly detection methods (iid/point-based, graph and time series) including active learning for anomaly detection/discovery, bayesian rule-mining, description for diversity/explanation/interpretability. Analysis of incorporating label feedback with ensemble and tree-based detectors. Includes adversarial attacks with Graph Convolutional Network.
active-learning adversarial-attacks anogan anomaly-detection autoencoder concept-drift ensemble-learning explaination gan generative-adversarial-network graph-convolutional-networks interpretability lstm nettack rnn streaming time-series timeseries trees unsuperivsed
Last synced: 16 Mar 2025
https://github.com/thunlp/openattack
An Open-Source Package for Textual Adversarial Attack.
adversarial-attacks adversarial-example natural-language-processing nlp pytorch
Last synced: 08 Apr 2025
https://github.com/thunlp/OpenAttack
An Open-Source Package for Textual Adversarial Attack.
adversarial-attacks adversarial-example natural-language-processing nlp pytorch
Last synced: 30 Mar 2025
https://github.com/hendrycks/natural-adv-examples
A Harder ImageNet Test Set (CVPR 2021)
adversarial-attacks adversarial-example domain-generalization imagenet ml-safety robustness
Last synced: 04 Apr 2025
https://github.com/MadryLab/photoguard
Raising the Cost of Malicious AI-Powered Image Editing
adversarial-attacks adversarial-examples computer-vision deep-learning deepfakes robustness stable-diffusion
Last synced: 03 Apr 2025
https://github.com/thu-ml/ares
A Python library for adversarial machine learning focusing on benchmarking adversarial robustness.
adversarial-attacks adversarial-machine-learning adversarial-robustness benchmark-framework bim boundary deepfool distillation evolutionary fgsm hgd mi-fgsm mmlda nes pca spsa
Last synced: 04 Apr 2025
https://github.com/jind11/TextFooler
A Model for Natural Language Attack on Text Classification and Inference
adversarial-attacks bert bert-model natural-language-inference natural-language-processing text-classification
Last synced: 18 Apr 2025
https://github.com/deadbits/vigil-llm
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
adversarial-attacks adversarial-machine-learning large-language-models llm-security llmops prompt-injection security-tools yara-scanner
Last synced: 06 Apr 2025
https://github.com/koukyosyumei/aijack
Security and Privacy Risk Simulator for Machine Learning (arXiv:2312.17667)
adversarial-attacks adversarial-examples adversarial-machine-learning dbms deep-learning differential-privacy evasion-attack federated-learning homomorphic-encryption k-anonymity machine-learning membership-inference model-inversion-attacks paillier paillier-cryptosystem poisoning-attacks privacy security
Last synced: 04 Apr 2025
https://github.com/HuntDownProject/HEDnsExtractor
A suite for hunting suspicious targets, expose domains and phishing discovery
adversarial-attacks malware-analysis phishing-detection threat-hunting threat-intelligence threatintel
Last synced: 09 Dec 2024
https://github.com/hbaniecki/adversarial-explainable-ai
💡 Adversarial attacks on explanations and how to defend them
adversarial adversarial-attacks adversarial-examples adversarial-machine-learning attacks counterfactual deep defense evaluation explainability explainable-ai iml interpretability interpretable interpretable-machine-learning model responsible-ai robustness security xai
Last synced: 25 Mar 2025
https://github.com/pumpbin/pumpbin
🎃 PumpBin is an Implant Generation Platform.
adversarial-attacks application bypass-antivirus c2 command-and-control cybersecurity iced implant linux macos offensive-security redteam rust security security-tools shellcode shellcode-injection shellcode-injector shellcode-loader windows
Last synced: 06 Apr 2025
https://github.com/chandlerbang/pro-gnn
Implementation of the KDD 2020 paper "Graph Structure Learning for Robust Graph Neural Networks"
adversarial-attacks attack-defense deep-learning defense graph-mining graph-neural-networks graph-structure-recovery machine-learning pytorch semi-supervised-learning
Last synced: 09 Apr 2025
https://github.com/ChandlerBang/Pro-GNN
Implementation of the KDD 2020 paper "Graph Structure Learning for Robust Graph Neural Networks"
adversarial-attacks attack-defense deep-learning defense graph-mining graph-neural-networks graph-structure-recovery machine-learning pytorch semi-supervised-learning
Last synced: 28 Nov 2024
https://github.com/ain-soph/trojanzoo
TrojanZoo provides a universal pytorch platform to conduct security researches (especially backdoor attacks/defenses) of image classification in deep learning.
adversarial-attacks backdoor-attacks deep-learning image-classification pytorch
Last synced: 18 Feb 2025
https://github.com/agencyenterprise/promptinject
PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to adversarial prompt attacks. 🏆 Best Paper Awards @ NeurIPS ML Safety Workshop 2022
adversarial-attacks agi agi-alignment ai-alignment ai-safety chain-of-thought gpt-3 language-models large-language-models machine-learning ml-safety prompt-engineering
Last synced: 05 Apr 2025
https://github.com/agencyenterprise/PromptInject
PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to adversarial prompt attacks. 🏆 Best Paper Awards @ NeurIPS ML Safety Workshop 2022
adversarial-attacks agi agi-alignment ai-alignment ai-safety chain-of-thought gpt-3 language-models large-language-models machine-learning ml-safety prompt-engineering
Last synced: 28 Mar 2025
https://github.com/natanielruiz/disrupting-deepfakes
🔥🔥Defending Against Deepfakes Using Adversarial Attacks on Conditional Image Translation Networks
adversarial-attacks computer-vision deep-learning deepfake-detection deepfakes defending defending-deepfakes disrupting-deepfakes face-swap faceswap fake-news machine-learning
Last synced: 18 Feb 2025
https://github.com/automorphic-ai/aegis
Self-hardening firewall for large language models
adversarial-attacks large-language-models llmops prompt-injection security
Last synced: 28 Mar 2025
https://github.com/VinAIResearch/Anti-DreamBooth
Anti-DreamBooth: Protecting users from personalized text-to-image synthesis (ICCV 2023)
adversarial-attacks dreambooth personalization stable-diffusion text-to-image
Last synced: 27 Mar 2025
https://github.com/ryderling/DEEPSEC
DEEPSEC: A Uniform Platform for Security Analysis of Deep Learning Model
adversarial-attacks adversarial-examples deep-leaning defenses
Last synced: 20 Apr 2025
https://github.com/The-Z-Labs/bof-launcher
Beacon Object File (BOF) launcher - library for executing BOF files in C/C++/Zig applications
adversarial-attacks beacon beaconobjectfile bof cobalt cobalt-strike coff cybersecurity elf execution in-memory penetration-testing-tools post-exploitation red-team security-tools
Last synced: 05 Apr 2025
https://github.com/bosch-aisecurity-aishield/watchtower
AIShield Watchtower: Dive Deep into AI's Secrets! 🔍 Open-source tool by AIShield for AI model insights & vulnerability scans. Secure your AI supply chain today! ⚙️🛡️
adversarial-attacks aisec keras mlops mlsecops pytorch safetensors scikit-learn security security-automation security-tools supply-chain tensorflow vulnerability-scanners
Last synced: 18 Apr 2025
https://github.com/the-z-labs/bof-launcher
Beacon Object File (BOF) launcher - library for executing BOF files in C/C++/Zig applications
adversarial-attacks beacon beaconobjectfile bof cobalt cobalt-strike coff cybersecurity elf execution in-memory penetration-testing-tools post-exploitation red-team security-tools
Last synced: 04 Apr 2025
https://github.com/harry24k/pgd-pytorch
A pytorch implementation of "Towards Deep Learning Models Resistant to Adversarial Attacks"
adversarial-attacks deep-learning pytorch
Last synced: 10 Apr 2025
https://github.com/safreita1/TIGER
Python toolbox to evaluate graph vulnerability and robustness (CIKM 2021)
adversarial-attacks attack cascading-failures data-mining data-science defense diffusion epidemics graph graph-attack graph-mining machine-learning netshield network-attack networks robustness simulation vulnerability
Last synced: 12 Nov 2024
https://github.com/PKU-YuanGroup/Hallucination-Attack
Attack to induce LLMs within hallucinations
adversarial-attacks ai-safety deep-learning hallucinations llm llm-safety machine-learning nlp
Last synced: 29 Dec 2024
https://github.com/OmidPoursaeed/Generative_Adversarial_Perturbations
Generative Adversarial Perturbations (CVPR 2018)
adversarial-attacks adversarial-example cvpr cvpr2018 deep-learning generative-model
Last synced: 02 Dec 2024
https://github.com/declare-lab/dialogue-understanding
This repository contains PyTorch implementation for the baseline models from the paper Utterance-level Dialogue Understanding: An Empirical Study
adversarial-attacks bert bert-embeddings conversational-agents conversational-ai dialogue dialogue-act dialogue-systems dialogue-understanding emotion-analysis emotion-recognition emotion-recognition-in-conversation explainable-ai lstm natural-language-processing natural-language-understanding pretrained-models pytorch sentiment-analysis utterance-level-dialogue
Last synced: 14 Apr 2025
https://github.com/git-disl/TOG
Real-time object detection is one of the key applications of deep neural networks (DNNs) for real-world mission-critical systems. While DNN-powered object detection systems celebrate many life-enriching opportunities, they also open doors for misuse and abuse. This project presents a suite of adversarial objectness gradient attacks, coined as TOG, which can cause the state-of-the-art deep object detection networks to suffer from untargeted random attacks or even targeted attacks with three types of specificity: (1) object-vanishing, (2) object-fabrication, and (3) object-mislabeling. Apart from tailoring an adversarial perturbation for each input image, we further demonstrate TOG as a universal attack, which trains a single adversarial perturbation that can be generalized to effectively craft an unseen input with a negligible attack time cost. Also, we apply TOG as an adversarial patch attack, a form of physical attacks, showing its ability to optimize a visually confined patch filled with malicious patterns, deceiving well-trained object detectors to misbehave purposefully.
adversarial-attacks faster-rcnn object-detection ssd yolov3
Last synced: 21 Apr 2025
https://github.com/gmh14/RobNets
[CVPR 2020] When NAS Meets Robustness: In Search of Robust Architectures against Adversarial Attacks
adversarial-attacks adversarial-examples deep-learning-architectures neural-architecture-search robustness
Last synced: 18 Nov 2024
https://github.com/safellama/plexiglass
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
adversarial-attacks adversarial-machine-learning cybersecurity deep-learning deep-neural-networks machine-learning security
Last synced: 18 Apr 2025
https://github.com/kortex-labs/plexiglass
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
adversarial-attacks adversarial-machine-learning cybersecurity deep-learning deep-neural-networks machine-learning security
Last synced: 21 Feb 2025
https://github.com/pku-yuangroup/hallucination-attack
Attack to induce LLMs within hallucinations
adversarial-attacks ai-safety deep-learning hallucinations llm llm-safety machine-learning nlp
Last synced: 10 Nov 2024
https://github.com/thudm/grb
Graph Robustness Benchmark: A scalable, unified, modular, and reproducible benchmark for evaluating the adversarial robustness of Graph Machine Learning.
adversarial-attacks deep-learning graph-neural-networks machine-learning pytorch
Last synced: 19 Dec 2024
https://github.com/Eaphan/Robust3DOD
A Comprehensive Study of the Robustness for LiDAR-based 3D Object Detectors against Adversarial Attacks
3d-object-detection adversarial-attacks
Last synced: 20 Mar 2025
https://github.com/shawnxyang/face-robustness-benchmark
An adversarial robustness evaluation library on face recognition.
adversarial-attacks face-recognition machine-learning robustness
Last synced: 23 Nov 2024
https://github.com/dipanjans/adversarial-learning-robustness
Contains materials for workshops pertaining to adversarial robustness in deep learning.
adversarial-attacks adversarial-learning computer-vision deep-learning python tensorflow
Last synced: 08 Mar 2025
https://github.com/thunlp/sememepso-attack
Code and data of the ACL 2020 paper "Word-level Textual Adversarial Attacking as Combinatorial Optimization"
adversarial-attacks adversarial-examples nlp pso sememe
Last synced: 10 Nov 2024
https://github.com/edisonleeeee/greatx
A graph reliability toolbox based on PyTorch and PyTorch Geometric (PyG).
adversarial-attacks distribution-shift graph-convolutional-networks graph-neural-networks graph-reliability-toolbox inherent-noise pytorch pytorch-geometric
Last synced: 05 Apr 2025
https://github.com/392781/faceoff
Steps towards physical adversarial attacks on facial recognition
adversarial-attacks adversarial-machine-learning facenet-pytorch facial-recognition physical-adversarial-attacks
Last synced: 17 Mar 2025
https://github.com/hfawaz/ijcnn19attacks
Adversarial Attacks on Deep Neural Networks for Time Series Classification
adversarial-attacks adversarial-examples deep-learning deep-neural-networks time-series-classification
Last synced: 09 Apr 2025
https://github.com/moohax/Proof-Pudding
Copy cat model for Proofpoint
adversarial-attacks adversarial-machine-learning offensive-machine-learning proofpoint
Last synced: 12 Nov 2024
https://github.com/jinzhuoran/RWKU
RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models. NeurIPS 2024
adversarial-attacks benchmark evaluation-framework forgetting large-language-models membership-inference-attack natural-language-processing privacy-protection right-to-be-forgotten unlearning
Last synced: 24 Mar 2025
https://github.com/dmitryryumin/wacv-2024-papers
WACV 2024 Papers: Discover cutting-edge research from WACV 2024, the leading computer vision conference. Stay updated on the latest in computer vision and deep learning, with code included. ⭐ support visual intelligence development!
3d-computer-vision 3d-sensor adversarial-attacks autonomous-driving biometrics computer-vision datasets face-recognition generative-models gesture-recognition image-recognition image-understanding low-level machine-learning robotics video-recognition vision-transformer visualization wacv wacv2024
Last synced: 12 Apr 2025
https://github.com/jinzhuoran/rwku
RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models. NeurIPS 2024
adversarial-attacks benchmark evaluation-framework forgetting large-language-models membership-inference-attack natural-language-processing privacy-protection right-to-be-forgotten unlearning
Last synced: 02 Apr 2025
https://github.com/ai4ce/flat
[ICCV2021 Oral] Fooling LiDAR by Attacking GPS Trajectory
3d-object-detection 3d-perception adversarial-attacks ai-safety autonomous-driving deep-learning gnss lidar point-cloud robotics trustworthy-ai trustworthy-machine-learning
Last synced: 11 Apr 2025
https://github.com/harry24k/fgsm-pytorch
A pytorch implementation of "Explaining and harnessing adversarial examples"
adversarial-attacks deep-learning pytorch
Last synced: 10 Apr 2025
https://github.com/snakers4/msc-2018-final
adversarial-attacks fgsm fgvm python3 pytorch
Last synced: 31 Mar 2025
https://github.com/pmichel31415/teapot-nlp
Tool for Evaluating Adversarial Perturbations on Text
adversarial-attacks natural-language-processing
Last synced: 14 Apr 2025
https://github.com/bhheo/bss_distillation
Knowledge Distillation with Adversarial Samples Supporting Decision Boundary (AAAI 2019)
adversarial-attacks adversarial-samples image-classification knowledge-distillation network-compression teacher-student-learning
Last synced: 12 Nov 2024
https://github.com/ermongroup/generative_adversary
Code for the unrestricted adversarial examples paper (NeurIPS 2018)
adversarial-attacks adversarial-examples neurips2018
Last synced: 18 Nov 2024
https://github.com/chandlerbang/simp-gcn
Implementation of the WSDM 2021 paper "Node Similarity Preserving Graph Convolutional Networks"
adversarial-attacks deep-learning graph-convolutional-networks graph-mining graph-neural-networks robustness
Last synced: 08 Apr 2025
https://github.com/anuragranj/flowattack
Attacking Optical Flow (ICCV 2019)
adversarial-attacks deep-learning optical-flow
Last synced: 05 Apr 2025
https://github.com/searchableai/kitanaqa
KitanaQA: Adversarial training and data augmentation for neural question-answering models
adversarial-attacks adversarial-training bert data-augmentation ml-automation natural-language-processing nlp pytorch question-answering transformer
Last synced: 10 Apr 2025
https://github.com/mims-harvard/GNNGuard
Defending graph neural networks against adversarial attacks (NeurIPS 2020)
adversarial-attacks deep-learning graph-convolutional-networks graph-neural-networks robust-learning
Last synced: 28 Nov 2024
https://github.com/Harry24k/CW-pytorch
A pytorch implementation of "Towards Evaluating the Robustness of Neural Networks"
adversarial-attacks deep-learning pytorch
Last synced: 03 Apr 2025
https://github.com/harry24k/cw-pytorch
A pytorch implementation of "Towards Evaluating the Robustness of Neural Networks"
adversarial-attacks deep-learning pytorch
Last synced: 10 Apr 2025
https://github.com/chandlerbang/gtrans
[ICLR'23] Implementation of "Empowering Graph Representation Learning with Test-Time Graph Transformation"
adversarial-attacks data-centric distribution-shift graph-neural-networks out-of-distribution robustness test-time-adaptation
Last synced: 08 Apr 2025
https://github.com/UCSC-VLAA/vllm-safety-benchmark
[ECCV 2024] Official PyTorch Implementation of "How Many Unicorns Are in This Image? A Safety Evaluation Benchmark for Vision LLMs"
adversarial-attacks benchmark datasets llm multimodal-llm robustness safety vision-language-model
Last synced: 02 Dec 2024
https://github.com/spencerwooo/torchattack
🛡 A curated list of adversarial attacks in PyTorch, with a focus on transferable black-box attacks.
adversarial-attacks adversarial-examples pytorch
Last synced: 14 Apr 2025
https://github.com/dsgiitr/adversarial_lab
Web-based Tool for visualisation and generation of adversarial examples by attacking ImageNet Models like VGG, AlexNet, ResNet etc.
adversarial-attacks computer-vision flask html-css-javascript imagenet machine-learning python pytorch visualization
Last synced: 15 Apr 2025
https://github.com/max-andr/provably-robust-boosting
Provably Robust Boosted Decision Stumps and Trees against Adversarial Attacks [NeurIPS 2019]
adversarial-attacks boosted-decision-stumps boosted-trees boosting provable-defense
Last synced: 27 Mar 2025
https://github.com/SamSamhuns/yolov5_adversarial
Generate adversarial patches against YOLOv5 🚀
adversarial-attacks adversarial-machine-learning object-detection
Last synced: 21 Apr 2025
https://github.com/ajhamdi/AdvPC
AdvPC: Transferable Adversarial Perturbations on 3D Point Clouds (ECCV 2020)
3d adversarial-attacks deep-learning eccv-2020 paper point-clouds pointnet tensorflow
Last synced: 20 Mar 2025
https://github.com/osu-nlp-group/amplegcg
AmpleGCG: Learning a Universal and Transferable Generator of Adversarial Attacks on Both Open and Closed LLM
adversarial-attacks gcg nlp safety
Last synced: 11 Nov 2024
https://github.com/ajhamdi/advpc
AdvPC: Transferable Adversarial Perturbations on 3D Point Clouds (ECCV 2020)
3d adversarial-attacks deep-learning eccv-2020 paper point-clouds pointnet tensorflow
Last synced: 23 Jan 2025
https://github.com/chawins/pal
PAL: Proxy-Guided Black-Box Attack on Large Language Models
adversarial-attacks jailbreak llm openai-api red-teaming
Last synced: 02 Dec 2024
https://github.com/xiaosen-wang/Adversarial-Examples-Paper
Paper list of Adversarial Examples
adversarial-attacks adversarial-examples
Last synced: 27 Mar 2025
https://github.com/daniel-liu-c0deb0t/adversarial-point-perturbations-on-3d-objects
New distributional and shape attacks on neural networks that process 3D point cloud data.
3d-point-clouds adversarial-attacks algorithms deep-learning matplotlib numba numpy pointnet python tensorflow
Last synced: 13 Apr 2025
https://github.com/Daniel-Liu-c0deb0t/Adversarial-point-perturbations-on-3D-objects
New distributional and shape attacks on neural networks that process 3D point cloud data.
3d-point-clouds adversarial-attacks algorithms deep-learning matplotlib numba numpy pointnet python tensorflow
Last synced: 20 Mar 2025
https://github.com/GiorgosKarantonis/Adversarial-Attacks-with-Relativistic-AdvGAN
Using relativism to improve GAN-based Adversarial Attacks. 🦾
adversarial-attacks adversarial-examples advgan artificial-intelligence gan generative-adversarial-networks machine-learning madrylab-challenge relativistic-gan rsgan
Last synced: 02 Dec 2024
https://github.com/harry24k/mair
Fantastic Robustness Measures: The Secrets of Robust Generalization [NeurIPS 2023]
adversarial-attacks adversarial-defenses adversarial-robustness adversarial-training
Last synced: 10 Apr 2025
https://github.com/paulescu/adversarial-machine-learning
Hands-on tutorial on adversarial examples 😈. With Streamlit app ❤️.
adversarial-attacks adversarial-machine-learning python
Last synced: 14 Apr 2025
https://github.com/imcaspar/audio-adv-papers
A curated list of awesome audio adversarial examples papers(with code & demo if available).
adversarial-attacks adversarial-examples
Last synced: 16 Feb 2025
https://github.com/yizhe-ang/detectron2-1
Implements Adversarial Examples for Semantic Segmentation and Object Detection, using PyTorch and Detectron2
adversarial-attacks adversarial-examples adversarial-machine-learning detectron detectron2 object-detection pytorch segmentation
Last synced: 01 Apr 2025
https://github.com/ChenWu98/agent-attack
[Arxiv 2024] Adversarial attacks on multimodal agents
adversarial-attacks agent llm multimodal safety
Last synced: 02 Dec 2024
https://github.com/harry24k/catastrophic-overfitting
Understanding Catastrophic Overfitting in Single-step Adversarial Training [AAAI 2021]
adversarial-attacks adversarial-defense adversarial-training catastrophic-overfitting
Last synced: 10 Apr 2025
https://github.com/wssun/TiSE-CodeLM-Security
This repository provide the studies on the security of language models for code (CodeLMs).
adversarial-attacks adversarial-defense ai-security ai4se backdoor-attacks backdoor-defense code-intelligence language-model lm4code lm4se security
Last synced: 27 Dec 2024
https://github.com/Harry24k/catastrophic-overfitting
Understanding Catastrophic Overfitting in Single-step Adversarial Training [AAAI 2021]
adversarial-attacks adversarial-defense adversarial-training catastrophic-overfitting
Last synced: 15 Nov 2024