Fact-checked by Grok 2 weeks ago
References
-
[1]
[PDF] Adversarial Machine Learning - NIST Technical Series PublicationsMar 20, 2025 · This NIST Trustworthy and Responsible AI report provides a taxonomy of concepts and defines terminology in the field of adversarial machine ...
-
[2]
[1312.6199] Intriguing properties of neural networks - arXivDec 21, 2013 · In this paper we report two such properties. First, we find that there is no distinction between individual high level units and random linear combinations of ...
-
[3]
[1412.6572] Explaining and Harnessing Adversarial Examples - arXivDec 20, 2014 · Explaining and Harnessing Adversarial Examples. Authors:Ian J. Goodfellow, Jonathon Shlens, Christian Szegedy.
-
[4]
Attacks in Adversarial Machine Learning: A Systematic Survey from ...Feb 19, 2023 · Abstract:Adversarial machine learning (AML) studies the adversarial phenomenon of machine learning, which may make inconsistent or unexpected ...
-
[5]
[PDF] A Comprehensive Review of Adversarial Attacks on Machine LearningDec 11, 2023 · These attacks involve crafting malicious inputs that can deceive a model into making incorrect predictions.<|separator|>
-
[6]
from Phenomena to Black-Box Attacks using Adversarial SamplesMay 24, 2016 · Abstract page for arXiv paper 1605.07277: Transferability in Machine Learning: from Phenomena to Black-Box Attacks using Adversarial ...Missing: date | Show results with:date
-
[7]
[1608.04644] Towards Evaluating the Robustness of Neural NetworksAug 16, 2016 · Title:Towards Evaluating the Robustness of Neural Networks. Authors:Nicholas Carlini, David Wagner ... attacks' ability to find adversarial ...
-
[8]
Towards Deep Learning Models Resistant to Adversarial AttacksJun 19, 2017 · We study the adversarial robustness of neural networks through the lens of robust optimization. This approach provides us with a broad and unifying view.
-
[9]
A Systematic Evaluation of Prompt Injection and Jailbreak ... - arXivMay 7, 2025 · Our experiments evaluated over 1,400 adversarial prompts across four LLMs: GPT-4, Claude 2, Mistral 7B, and Vicuna. We analyze results along ...
-
[10]
[PDF] A Comprehensive Study of Jailbreak Attack versus Defense for ...Aug 11, 2024 · Jailbreak attacks use prompts to bypass safety measures in LLMs, producing harmful content. This study analyzes 9 attack and 7 defense ...
-
[11]
[PDF] arXiv:2402.06363v2 [cs.CR] 25 Sep 2024Sep 25, 2024 · Prompt injection attacks pose a major challenge for devel- oping secure LLM-integrated applications, as they typically need to process much data ...
-
[12]
BadFU: Backdoor Federated Learning through Adversarial Machine ...Aug 21, 2025 · Specifically, we propose BadFU, an attack strategy where a malicious client uses both backdoor and camouflage samples to train the global model ...
-
[13]
[PDF] FedBAP: Backdoor Defense via Benign Adversarial Perturbation in ...Jul 26, 2025 · Federated Learning (FL) enables collaborative model training while preserving data privacy, but it is highly vulnerable to backdoor attacks.
-
[14]
Defending backdoor attacks in federated learning via adversarial ...This paper proposes ADFL, a novel adversarial distillation-based backdoor defense scheme for federated learning.
-
[15]
AI 100-2 E2023, Adversarial Machine Learning: A Taxonomy and ...Jan 4, 2024 · This NIST Trustworthy and Responsible AI report develops a taxonomy of concepts and defines terminology in the field of adversarial machine learning (AML).
-
[16]
AI 100-2 E2025, Adversarial Machine Learning: A Taxonomy and ...This NIST Trustworthy and Responsible AI report provides a taxonomy of concepts and defines terminology in the field of adversarial machine learning (AML).
-
[17]
AdvML 2024 - New Frontiers in Adversarial Machine LearningDec 14, 2024 · Join us at AdvML-Frontiers'24 for a comprehensive exploration of adversarial learning at the intersection with cutting-edge multimodal technologies.
-
[18]
Universal Adversarial Attack on Multimodal Aligned LLMs - arXivJun 4, 2025 · We propose a universal adversarial attack on multimodal Large Language Models (LLMs) that leverages a single optimized image to override ...
-
[19]
NIST AI 100-2 | Adversarial Machine Learning Taxonomy | CSRCNIST AI 100-2 targets this issue and offers voluntary guidance relative to identifying, addressing, and managing the risks associated with adversarial machine ...
-
[20]
[PDF] Adversarial Machine Learning: A Taxonomy and Terminology of ...Jan 2, 2024 · This NIST Trustworthy and Responsible AI report develops a taxonomy of concepts and defines terminology in the field of adversarial machine ...
-
[21]
[PDF] Simple Black-box Adversarial AttacksUnder the white-box threat model, the classifier h is provided to the adversary. In this scenario, a powerful attack strategy is to perform gradient descent on ...
- [22]
-
[23]
[PDF] Adversarial Machine Learning∗ - People @EECSOct 21, 2011 · Cost models of the adversary also led to a theory for query-based near-optimal evasion of classifiers first presented by Lowd and Meek, in which ...
-
[24]
[PDF] A Survey of Game Theoretic Approaches for Adversarial Machine ...The influence dimension specifies two types of adversarial attacks, causative and exploratory (also known as probing), illustrated in figure 2. In causative ...Missing: explorative | Show results with:explorative
-
[25]
A taxonomy and survey of attacks against machine learningThis paper presents a taxonomy and survey of attacks against systems that use machine learning. It organizes the body of knowledge in adversarial machine ...
-
[26]
Machine Learning Models Have a Supply Chain Problem - arXivMay 28, 2025 · In this paper, we argue that the current ecosystem for open ML models contains significant supply-chain risks, some of which have been exploited ...3 Model Transparency · 5 Integrity For Ml Models · 6 Dataset Verifiability
- [27]
-
[28]
[PDF] Exploring the Limits of Model-Targeted Indiscriminate Data ...Figure 10: We visualize some poisoned images generated by the GC attack on the CIFAR-10 dataset. The first row shows the clean samples, the second row shows ...
-
[29]
[PDF] Detecting and Preventing Data Poisoning Attacks on AI Models - arXivExperimental results indicate that data poisoning significantly degrades model performance, reducing classification accuracy by up to 27% in image recognition ...
-
[30]
[PDF] How To Backdoor Federated LearningWe show that this makes federated learning vulnerable to a model-poisoning attack that is signifi- cantly more powerful than poisoning attacks that target only ...
-
[31]
Exploring Backdoor Attacks against Personalized Federated LearningJan 22, 2025 · Data heterogeneity and backdoor attacks rank among the most significant challenges facing federated learning (FL). For data heterogeneity, ...
-
[32]
[PDF] Byzantine-Robust Distributed Learning: Towards Optimal Statistical ...In this paper, we develop distributed optimiza- tion algorithms that are provably robust against. Byzantine failures—arbitrary and potentially ad- versarial ...
-
[33]
[PDF] Byzantine Fault Tolerance in Distributed Machine Learning - arXivDec 4, 2022 · Byzantine network failures for distributed convex and non-convex learning tasks. ... attacks, single point of failure, data privacy, and existing ...
-
[34]
[PDF] Black-box Adversarial Attacks with Limited Queries and InformationThis paper defines query-limited, partial-information, and label-only settings, where attackers have limited queries, partial information, or only the top ...
-
[35]
[PDF] Towards More Practical Threat Models in Artificial Intelligence SecurityIn model stealing, the attacker has black- box access to an ML model and copies its functionality with- out consent of the model's owner [71] and thus harms con ...
-
[36]
[PDF] From AI Vulnerabilities to AI Security Incident Reporting• 2 Model Stealing. • 2 Poisoning. • 4 Privacy. • 10 Cybersecurity. 27. Found Incidents. Preventable following. Security Best Practices! (2) Malicious intent.
-
[37]
RobustBench: a standardized adversarial robustness benchmarkOct 19, 2020 · Our goal is to establish a standardized benchmark of adversarial robustness, which as accurately as possible reflects the robustness of the considered models.Missing: CNN PGD drop
-
[38]
RobustBench: Adversarial robustness benchmarkA standardized benchmark for adversarial robustness. The goal of RobustBench is to systematically track the real progress in adversarial robustness.Missing: CNN PGD drop
-
[39]
Chapter 2 - linear modelsUnderstanding the linear case provides important insights into the theory and practice of adversarial robustness, and also provides connections to more ...Missing: neural | Show results with:neural
-
[40]
[PDF] Adversarial Robustness of Deep Neural Networks - arXivFurthermore, neural networks themselves are often vulnerable to adversarial attacks. For those reasons, there is a high demand for trustworthy and rigorous ...
-
[41]
[PDF] Adversarial Examples are not Bugs, they are FeaturesOne of the most intriguing properties of adversarial examples is that they transfer across models with different architectures and independently sampled ...<|control11|><|separator|>
-
[42]
[PDF] Why Do Adversarial Attacks Transfer? Explaining ... - USENIXAug 14, 2019 · We give a formal definition of transferability of evasion and poisoning attacks, and an upper bound on a transfer attack's success.
-
[43]
[PDF] A Survey on Transferability of Adversarial Examples Across Deep ...May 2, 2024 · Adversarial examples are specially crafted inputs that lead machine learning models to make incorrect predictions. These inputs are impercep- ...
-
[44]
Robust Deep Reinforcement Learning with Adversarial Attacks - arXivThis paper proposes adversarial attacks for Reinforcement Learning (RL) and then improves the robustness of Deep Reinforcement Learning algorithms (DRL) to ...
-
[45]
[PDF] Robust Deep Reinforcement Learning against Adversarial ...A deep reinforcement learning (DRL) agent observes its states through observa- tions, which may contain natural measurement errors or adversarial noises.
-
[46]
A study of natural robustness of deep reinforcement learning ...Analyzing the robustness of DRL algorithms to adversarial attacks is an important prerequisite to enabling the widespread adoption of DRL algorithms. Common ...
-
[47]
[PDF] Mission Impossible: A Statistical Perspective on Jailbreaking LLMsAttack success rate numbers are shown in Table 4, with each model's MT-Bench scores shown in the brackets. Clearly, almost all safe prefixes lead to better ...<|separator|>
-
[48]
Adversarial Examples Are Not Bugs, They Are Features - arXivMay 6, 2019 · We demonstrate that adversarial examples can be directly attributed to the presence of non-robust features: features derived from patterns in the data ...
- [49]
-
[50]
[PDF] Adversarial Examples Are Not Easily Detected - Nicholas CarliniWe use the L2 attack algorithm of Carlini and Wagner [8] to generate targeted adversarial examples, as it is superior to other published attacks. At a high ...
-
[51]
White-box and Black-box Attacks for Transfer Learning - ADSEmpirical results show that the adversarial examples are more transferable when fine-tuning is used than they are when the two networks are trained ...
-
[52]
[PDF] Enhancing the Transferability of Adversarial Examples with Random ...Compared to the state-of-the-art transferable attacks, our attacks improve the black-box attack success rate by 2.9% against normally trained mod- els, 4.7% ...
- [53]
-
[54]
[PDF] Bridging the Gap Between Adversarial ML Research and PracticeOur analysis clearly indicates that real adversaries do attempt to evade anti-phishing ML systems that use image classification, and do so with some degree of ...
- [55]
-
[56]
Theoretically Principled Trade-off between Robustness and AccuracyJan 24, 2019 · We identify a trade-off between robustness and accuracy that serves as a guiding principle in the design of defenses against adversarial examples.
-
[57]
DNN-Oriented JPEG Compression Against Adversarial ExamplesMar 14, 2018 · We propose a JPEG-based defensive compression framework, namely "feature distillation", to effectively rectify adversarial examples without impacting ...Missing: effectiveness | Show results with:effectiveness
-
[58]
Full article: Adversarial Machine Learning: Bayesian PerspectivesWe demonstrate how the Bayesian approach allows us to explicitly model our uncertainty about the opponent's beliefs and interests, relaxing unrealistic ...
-
[59]
Automatic Evasion of Machine Learning-Based Network Intrusion ...Final results show that the proposed strategy effectively evades seven typical ML-based IDSs and one SOTA DL-based IDS with an average success rate of over < ...
-
[60]
Adversarial attack vulnerability of medical image analysis systemsIn this paper, we study previously unexplored factors affecting adversarial attack vulnerability of deep learning MedIA systems in three medical domains.Missing: incidents | Show results with:incidents
-
[61]
Adversarial AI threatens our financial services. We need a response.Jan 22, 2025 · Model theft. AI models are high value intellectual property: crown jewels to protect from theft. However, a technique known as 'model ...
- [62]
-
[63]
[PDF] Stealing Machine Learning Models via Prediction APIs - USENIXAug 10, 2016 · On Google's plat- form for example, an extraction attack would cost less than $0. ... model extraction attacks that could subvert model mon-.
-
[64]
How to Keep a Model Stealing Adversary Busy? - CleverHans LabApr 21, 2022 · A proactive defense using proof-of-work (PoW) puzzles, with difficulty calibrated to query leakage, increases the cost of model extraction by ...
-
[65]
Adversarial Machine Learning - Joint Air Power Competence CentreNowadays, the lack of robustness of these systems can no longer be ignored; many of them have proven to be highly vulnerable to intentional adversarial attacks ...
-
[66]
[PDF] Policy Teaching in Reinforcement Learning via Environment ...We study a security threat to reinforcement learning where an attacker poisons the learning environment to force the agent into executing a target policy ...
-
[67]
Vulnerability-Aware Poisoning Mechanism for Online RL with ... - arXivSep 2, 2020 · We propose a strategic poisoning algorithm called Vulnerability-Aware Adversarial Critic Poison (VA2C-P), which works for most policy-based deep RL agents.
-
[68]
Adversarial Machine Learning in Industry: A Systematic Literature ...This literature study reviews studies in the area of AML in the context of industry, measuring and analyzing each study's rigor and relevance scores.
-
[69]
A Systematic Survey of Model Extraction Attacks and Defenses - arXivAug 20, 2025 · Zhou et al. (2024) propose an inversion-guided defense that detects potential model stealing attacks by analyzing the invertibility of the ...
-
[70]
Adversarial ML Problems Are Getting Harder to Solve and to EvaluateThe field of adversarial ML studies problems that are (1) less clearly defined, (2) harder to solve, and (3) even more challenging to evaluate.
-
[71]
Feasibility of adversarial attacks against machine learning modelsDec 11, 2024 · This provides a clearer understanding of how to make models more resilient in real-world situations, where attackers face more limitations.
-
[72]
Evaluating the Evaluators: Trust in Adversarial Robustness TestsJul 4, 2025 · We present AttackBench, a benchmark framework developed to assess the effectiveness of gradient-based attacks under standardized and reproducible conditions.
-
[73]
[PDF] Evaluating the Evaluators: Trust in Adversarial Robustness TestsJun 24, 2025 · Together, these inconsistencies introduce variance that can severely distort robustness assessments, hinder reproducibility, and create a false ...
-
[74]
Leakage and the reproducibility crisis in machine-learning-based ...Sep 8, 2023 · We surveyed a variety of research that uses ML and found that data leakage affects at least 294 studies across 17 fields, leading to overoptimistic findings.
-
[75]
Adversarial AI: Coming of age or overhyped?Sep 1, 2023 · This article explores developments in adversarial artificial intelligence (AAI) and machine learning, examining recent research, practical realities<|separator|>