<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"><channel><title>Adversarial ML</title><description>Adversarial ML coverage for engineers shipping ML systems. Membership inference, model extraction, evasion attacks, training-data extraction, backdoors — focused on what&apos;s exploitable against deployed models and what defenders can actually do about it. PoCs against open models, behavioral analysis for closed ones.</description><link>https://adversarialml.dev/</link><language>en</language><item><title>Data Poisoning and Backdoor Attacks on Foundation Models</title><link>https://adversarialml.dev/posts/data-poisoning-backdoor-attacks/</link><guid isPermaLink="true">https://adversarialml.dev/posts/data-poisoning-backdoor-attacks/</guid><description>Training data manipulation, backdoor triggers, and Trojan attacks against large-scale models. What the threat model actually requires and where the defenses are in 2026.</description><pubDate>Sun, 10 May 2026 00:00:00 GMT</pubDate><category>data-poisoning</category><category>backdoor-attacks</category><category>trojan-ml</category><category>adversarial-ml</category><category>ml-security</category><category>foundation-models</category><author>Adversarial ML Editorial</author></item><item><title>Evasion Attacks on Image Classifiers: FGSM, PGD, and C&amp;W</title><link>https://adversarialml.dev/posts/evasion-attacks-fgsm-pgd-cw/</link><guid isPermaLink="true">https://adversarialml.dev/posts/evasion-attacks-fgsm-pgd-cw/</guid><description>The three foundational gradient-based evasion attacks, what each one actually optimizes, and what the benchmark numbers mean when you&apos;re evaluating a defense.</description><pubDate>Sun, 10 May 2026 00:00:00 GMT</pubDate><category>evasion-attacks</category><category>fgsm</category><category>pgd</category><category>carlini-wagner</category><category>adversarial-examples</category><category>adversarial-ml</category><category>image-classifiers</category><author>Adversarial ML Editorial</author></item><item><title>Adversarial Robustness in NLP: Why Text Attacks Are Different</title><link>https://adversarialml.dev/posts/adversarial-robustness-nlp-text/</link><guid isPermaLink="true">https://adversarialml.dev/posts/adversarial-robustness-nlp-text/</guid><description>Discrete input spaces, semantic constraints, and human-perceptibility rules change what counts as an adversarial example in text. The attacks are harder to define and harder to defend.</description><pubDate>Sun, 10 May 2026 00:00:00 GMT</pubDate><category>adversarial-nlp</category><category>text-attacks</category><category>robustness</category><category>nlp</category><category>adversarial-ml</category><category>ml-security</category><category>transformers</category><author>Adversarial ML Editorial</author></item><item><title>Adversarial Transferability: Why Black-Box Attacks Work at All</title><link>https://adversarialml.dev/posts/transferability-black-box-attacks/</link><guid isPermaLink="true">https://adversarialml.dev/posts/transferability-black-box-attacks/</guid><description>Adversarial examples transfer across models with different architectures and training sets. Understanding why changes what you think defenses need to accomplish.</description><pubDate>Sun, 10 May 2026 00:00:00 GMT</pubDate><category>transferability</category><category>black-box-attacks</category><category>adversarial-examples</category><category>evasion</category><category>adversarial-ml</category><category>ml-security</category><author>Adversarial ML Editorial</author></item><item><title>Model Inversion Attacks: Reconstructing Training Data from Model Outputs</title><link>https://adversarialml.dev/posts/model-inversion-attacks/</link><guid isPermaLink="true">https://adversarialml.dev/posts/model-inversion-attacks/</guid><description>From Fredrikson&apos;s pharmacogenetics exploit to Geiping&apos;s gradient inversion, model inversion attacks recover private training data in ways most ML engineers don&apos;t expect.</description><pubDate>Sun, 10 May 2026 00:00:00 GMT</pubDate><category>model-inversion</category><category>privacy</category><category>gradient-inversion</category><category>training-data</category><category>adversarial-ml</category><category>federated-learning</category><author>Adversarial ML Editorial</author></item><item><title>Certified Robustness via Randomized Smoothing: What &apos;Certified&apos; Actually Guarantees</title><link>https://adversarialml.dev/posts/certified-robustness-randomized-smoothing/</link><guid isPermaLink="true">https://adversarialml.dev/posts/certified-robustness-randomized-smoothing/</guid><description>Randomized smoothing gives you a provable robustness radius. Understanding what that certificate means in practice — and where it breaks — is more useful than the headline number.</description><pubDate>Sat, 09 May 2026 00:00:00 GMT</pubDate><category>certified-robustness</category><category>randomized-smoothing</category><category>adversarial-defense</category><category>ml-security</category><category>formal-verification</category><author>Adversarial ML Editorial</author></item><item><title>Training Data Extraction from LLMs: The Carlini et al. Results and What They Mean</title><link>https://adversarialml.dev/posts/training-data-extraction-llms/</link><guid isPermaLink="true">https://adversarialml.dev/posts/training-data-extraction-llms/</guid><description>Carlini et al. demonstrated verbatim extraction of training data from GPT-2. The results have been widely misread. Here&apos;s what the paper actually shows, what makes data extractable, and what production mitigations work.</description><pubDate>Fri, 08 May 2026 00:00:00 GMT</pubDate><category>training-data-extraction</category><category>memorization</category><category>privacy</category><category>llm-security</category><category>gdpr</category><author>Adversarial ML Editorial</author></item><item><title>GCG-Class Adversarial Suffix Attacks: A 2026 Practitioner Primer</title><link>https://adversarialml.dev/posts/gcg-class-adversarial-suffix-2026/</link><guid isPermaLink="true">https://adversarialml.dev/posts/gcg-class-adversarial-suffix-2026/</guid><description>The math, the cost curve, and why optimization-based attacks are now within reach of solo practitioners. With reproducible setup and what defenders actually need to do.</description><pubDate>Thu, 07 May 2026 00:00:00 GMT</pubDate><category>adversarial-ml</category><category>gcg</category><category>optimization-attacks</category><category>red-team</category><category>alignment</category><author>Adversarial ML Editorial</author></item><item><title>Membership Inference Attacks: What Actually Works Against Production ML APIs</title><link>https://adversarialml.dev/posts/membership-inference-attacks/</link><guid isPermaLink="true">https://adversarialml.dev/posts/membership-inference-attacks/</guid><description>Shokri et al.&apos;s shadow-model attack is the canonical reference, but the gap between the paper&apos;s threat model and a real rate-limited API is wide. Here&apos;s what survives that gap.</description><pubDate>Thu, 07 May 2026 00:00:00 GMT</pubDate><category>membership-inference</category><category>privacy</category><category>ml-security</category><category>production-ml</category><category>red-team</category><author>Adversarial ML Editorial</author></item><item><title>Model Extraction via Query-Based Functional Stealing</title><link>https://adversarialml.dev/posts/model-extraction-attacks/</link><guid isPermaLink="true">https://adversarialml.dev/posts/model-extraction-attacks/</guid><description>Query-based model stealing attacks can recover a functionally equivalent model from API access alone. The economics matter more than the technique: here&apos;s when extraction is worth doing.</description><pubDate>Thu, 07 May 2026 00:00:00 GMT</pubDate><category>model-extraction</category><category>model-stealing</category><category>ml-security</category><category>adversarial-ml</category><category>api-security</category><author>Adversarial ML Editorial</author></item><item><title>What this site is for</title><link>https://adversarialml.dev/posts/welcome/</link><guid isPermaLink="true">https://adversarialml.dev/posts/welcome/</guid><description>Adversarial ML covers attacks against deployed ML systems and the defenses that hold up. Here&apos;s what we publish.</description><pubDate>Sun, 03 May 2026 00:00:00 GMT</pubDate><category>meta</category><author>Adversarial ML Editorial</author></item></channel></rss>