Accepted Papers

Below is the list of the accepted papers at SaTML 2026, organized by category. To learn more about the three categories of papers, please visit the Call for Papers.

Research Papers

Smudged Fingerprints: A Systematic Evaluation of the Robustness of AI Image Fingerprints
Kai Yao and Marc Juarez (University of Edinburgh)
Privacy Risks in Time Series Forecasting: User- and Record-Level Membership Inference
Nicolas Johansson, Tobias Olsson (Chalmers University of Technology), Daniel Nilsson, Johan Östman and Fazeleh Hoseini (AI Sweden)
Certifiably Robust RAG against Retrieval Corruption
Chong Xiang (NVIDIA), Tong Wu, Zexuan Zhong (Princeton University), David Wagner (University of California, Berkeley), Danqi Chen and Prateek Mittal (Princeton University)
FedSpy-LLM: Towards Scalable and Generalizable Data Reconstruction Attacks from Gradients on LLMs
Syed Irfan Ali Meerza (University of Tennessee Knoxville), Feiyi Wang (Oak Ridge National Laboratory) and Jian Liu (University of Georgia)
Training Set Reconstruction from Differentially Private Forests: How Effective is DP?
Alice Gorgé (École Polytechnique, Palaiseau), Julien Ferry (Polytechnique Montréal), Sébastien Gambs (Université du Québec à Montréal) and Thibaut Vidal (Polytechnique Montréal)
Evaluating Deep Unlearning in Large Language Models
Ruihan Wu (University of California, San Diego), Chhavi Yadav, Ruslan Salakhutdinov (CMU) and Kamalika Chaudhuri (University of California, San Diego)
Efficient and Scalable Implementation of Differentially Private Deep Learning without Shortcuts
Sebastian Rodriguez Beltran, Marlon Tobaben, Joonas Jälkö (University of Helsinki), Niki Loppi (NVIDIA) and Antti Honkela (University of Helsinki)
Defeating Prompt Injections by Design
Edoardo Debenedetti (ETH Zurich), Ilia Shumailov (Google DeepMind), Tianqi Fan (Google), Jamie Hayes (Google DeepMind), Nicholas Carlini (Anthropic), Daniel Fabian, Christoph Kern (Google), Chongyang Shi, Andreas Terzis (Google DeepMind) and Florian Tramèr (ETH Zurich)
ConCap: Practical Network Traffic Generation for (ML- and) Flow-based Intrusion Detection Systems
Miel Verkerken (Ghent University - imec), Laurens D'hooge (Ghent University - imec, Department of Information Technology, IDLab), Bruno Volckaert (IDLab-imec, Ghent University), Filip De Turck (Ghent University - imec) and Giovanni Apruzzese (University of Liechtenstein)
Beyond the TESSERACT: Trustworthy Dataset Curation for Sound Evaluations of Android Malware Classifiers
Theo Chow (King's College London, University College London), Mario D'Onghia (University College London), Lorenz Linhardt (Technische Universität Berlin, BIFOLD), Zeliang Kan (HiddenLayer, King's College London), Daniel Arp (Technische Universität Wien), Lorenzo Cavallaro and Fabio Pierazzi (University College London)
Targeting Alignment: Extracting Safety Classifiers of Aligned LLMs
Jean-Charles Noirot Ferrand, Yohan Beugin (University of Wisconsin-Madison), Eric Pauley (Virginia Tech), Ryan Sheatsley and Patrick McDaniel (University of Wisconsin-Madison)
Optimal Robust Recourse with $L^p$-Bounded Model Change
Phone Kyaw, Kshitij Kayastha and Shahin Jabbari (Drexel University)
Towards Zero Rotation and Beyond: Architecting Neural Networks for Fast Secure Inference with Homomorphic Encryption
Yifei Cai (Iowa State University), Yizhou Feng (Old Dominion University), Qiao Zhang (Shandong University), Chunsheng Xin (Iowa State University) and Hongyi Wu (University of Arizona)
Evaluating Black-Box Vulnerabilities with Wasserstein-Constrained Data Perturbations
Adriana Laurindo Monteiro (FGV - EMap) and Jean-Michel Loubes (Université Paul Sabatier)
Off-The-Shelf Image-to-Image Models Are All You Need To Defeat Image Protection Schemes
Xavier Pleimling, Sifat Muhammad Abdullah (Virginia Tech), Gunjan Balde (Indian Institute of Technology Kharagpur), Peng Gao (Virginia Tech), Mainack Mondal (Indian Institute of Technology Kharagpur), Murtuza Jadliwala (University of Texas at San Antonio) and Bimal Viswanath (Virginia Tech)
Membership Inference Attacks for Retrieval Based In-Context Learning for Document Question Answering
Tejas Kulkarni, Antti Koskela (Nokia Bell Labs) and Laith Zumot (Nokia)
Differentially Private Adaptation of Diffusion Models via Noisy Aggregated Embeddings
Pura Peetathawatchai (ETH Zurich), Wei-Ning Chen (Microsoft), Berivan Isik (Google), Sanmi Koyejo (Stanford University) and Albert No (Yonsei University)
Provably Safe Model Updates
Leo Elmecker-Plakolm (Imperial College London), Pierre Fasterling (EPFL), Philip Sosnin, Calvin Tsay and Matthew Wicker (Imperial College London)
Exact Unlearning of Finetuning Data via Model Merging at Scale
Kevin Kuo, Amrith Setlur, Kartik Srinivas, Aditi Raghunathan and Virginia Smith (Carnegie Mellon University)
Architectural Backdoors for Within-Batch Data Stealing and Model Inference Manipulation
Nicolas Küchler (ETH Zurich), Ivan Petrov, Conrad Grobler and Ilia Shumailov (Google DeepMind)
BinaryShield: Cross-Service Threat Intelligence in LLM Services using Privacy-Preserving Fingerprints
Waris Gill, Natalie Isak and Matthew Dressman (Microsoft)
Oblivious Exact (Un)Learning of Extremely Randomized Trees
Sofiane Azogagh, Zelma Aubin Birba, Sébastien Gambs and Marc-Olivier Killijian (Université du Québec à Montréal)
DeepLeak: Privacy Enhancing Hardening of Model Explanations Against Membership Leakage
Firas Ben Hmida, Zain Sbeih, Philemon Hailemariam and Birhanu Eshete (University of Michigan, Dearborn)
On the Fragility of Contribution Evaluation in Federated Learning
Balázs Pejó (EGroup), Marcell Frank (VIK - BME), Krisztian Varga, Peter Veliczky (TTK - BME) and Gergely Biczok (HUN-REN)
Stealthy Fake News and Lost Profits: Manipulating Headlines in LLM-Driven Algorithmic Trading
Advije Rizvani, Giovanni Apruzzese and Pavel Laskov (University of Liechtenstein)
“Org-Wide, We’re Not Ready": C-Level Lessons on Securing Generative AI Systems
Elnaz Rabieinejad Balagafsheh, Ali Dehghantanha (Cyber Science Lab, Canada Cyber Foundry, University of Guelph) and Fattane Zarrinkalam (College of Engineering, University of Guelph)
Reconstructing Training Data from Models Trained with Transfer Learning
Yakir Oz, Gilad Yehudai, Gal Vardi, Itai Antebi, Michal Irani and Niv Haim (Weizmann Institute of Science)
Private Blind Model Averaging – Distributed, Non-interactive, and Convergent
Moritz Kirschte, Sebastian Meiser (University of Lubeck), Saman Ardalan (UKSH Kiel) and Esfandiar Mohammadi (University of Lubeck)
Your Privacy Depends on Others: Collusion Vulnerabilities in Individual Differential Privacy
Johannes Kaiser (Technical University of Munich), Alexander Ziller (Technical Universtiy of Munich), Eleni Triantafillou (Google Deepmind), Daniel Rückert (Technical University of Munich, Imperial College London) and Georgios Kaissis (Technical University of Munich, Google Deepmind)
Counterfactual Training: Teaching Models Plausible and Actionable Explanations
Patrick Altmeyer, Aleksander Buszydlik, Arie van Deursen and Cynthia C. S. Liem (Delft University of Technology)
The Feature-Space Illusion: Exposing Practical Vulnerabilities in Blockchain GNN Fraud Detection
François Frankart, Thibault Simonetto, Maxime Cordy, Orestis Papageorgiou, Nadia Pocher and Gilbert Fridgen (University of Luxembourg)
Reasoning Introduces New Poisoning Attacks Yet Makes Them More Complicated
Hanna Foerster (University of Cambridge), Ilia Shumailov (Google Deepmind), Yiren Zhao (Imperial College London), Harsh Chaudhari (Northeastern University), Jamie Hayes (Google Deepmind), Robert Mullins (University of Cambridge) and Yarin Gal (University of Oxford)
Structured Command Hijacking against Embodied Artificial Intelligence with Text-based Controls
Luis Burbano, Diego Ortiz (University of California, Santa Cruz), Qi Sun (Johns Hopkins University), Siwei Yang, Haoqin Tu, Cihang Xie (University of California, Santa Cruz), Yinzhi Cao (Johns Hopkins University) and Alvaro A Cardenas (University of California, Santa Cruz)
Are Robust Fingerprints Adversarially Robust?
Anshul Nasery (University of Washington), Edoardo Contente (Sentient Research), Alkin Kaz, Pramod Viswanath (Princeton University) and Sewoong Oh (University of Washington)
One RNG to Rule Them All - How Randomness Becomes an Attack Vector in Machine Learning
Kotekar Annapoorna Prabhu, Andrew Gan and Zahra Ghodsi (Purdue University)
Safe But Not Robust: Security Evaluation of VLM by Jailbreaking MSTS
Wenxin Ding (University of Chicago), Cong Chen, Jean-Philippe Monteuuis and Jonathan Petit (Qualcomm)
Gauss-Newton Unlearning for the LLM Era
Lev McKinney, Anvith Thudi, Juhan Bae (University of Toronto), Tara Rezaei Kheirkhah (Massachusetts Institute of Technology), Nicolas Papernot, Sheila A. McIlraith and Roger Baker Grosse (University of Toronto)
A False Sense of Privacy: Evaluating Textual Data Sanitization Beyond Surface-level Privacy Leakage
Rui Xin (University of Washington), Niloofar Mireshghallah (CMU), Shuyue Stella Li, Michael Duan (University of Washington), Hyunwoo Kim (NVIDIA), Yejin Choi (Stanford), Yulia Tsvetkov, Sewoong Oh and Pang Wei Koh (University of Washington)
Temporal Misalignment Attacks against Multimodal Perception in Autonomous Driving
Md Hasan Shahriar, Md Mohaimin Al Barat (Virginia Tech), Harshavardhan Sundar (Amazon.com, Inc.), Ning Zhang (Washington University in St. Louis), Naren Ramakrishnan, Y. Thomas Hou and Wenjing Lou (Virginia Tech)
RobPI: Robust Private Inference against Malicious Client
Jiaqi Xue, Mengxin Zheng and Qian Lou (University of Central Florida)
Defending Against Prompt Injection with DataFilter
Yizhu Wang, Sizhe Chen (UC Berkeley), Raghad Alkhudair, Basel Alomair (KACST) and David Wagner (UC Berkeley)
Accelerating Targeted Hard-Label Adversarial Attacks in Low-Query Black-Box Settings
Arjhun Swaminathan and Mete Akgün (Eberhard Karls Universität Tübingen)
RobustBlack: Challenging Black-Box Adversarial Attacks on State-of-the-Art Defenses
Mohamed DJILANI (University of Luxembourg), Salah GHAMIZI (Luxembourg Institute of Health) and Maxime CORDY (University of Luxembourg)
On the Effectiveness of Membership Inference in Targeted Data Extraction from Large Language Models
Ali Al Sahili, Ali Chehab and Razan Tajeddine (American University of Beirut)
On the Robustness of Tabular Foundation Models: Test-Time Attacks and In-Context Defenses
Mohamed Djilani, Thibault Simonetto, Karim Tit, Florian Tambon (University of Luxembourg), Salah Ghamizi (Luxembourg Institute of Health), Maxime Cordy and Mike Papadakis (University of Luxembourg)
Protecting Facial Biometrics from Malicious Generative Editing via Latent Optimization
Fahad Shamshad, Hashmat Shadab Malik (Mohamed bin Zayed University of Artificial Intelligence, UAE), Muzammal Naseer (Khalifa University, UAE), Salman Khan (Mohamed bin Zayed University of Artificial Intelligence, UAE. The Australian National University) and Karthik Nandakumar (Mohamed bin Zayed University of Artificial Intelligence, UAE. Michigan State University, USA)
They’re Closer Than We Think: Tackling Near-OOD Problem
Shaurya Bhatnagar, Ishika Sharma, Ranjitha Prasad (Indraprastha Institute of Information Technology Delhi (IIIT-Delhi)), Vidya T (LightMetrics), Ramya Hebbalaguppe (TCS Research) and Ashish Sethi (LightMetrics)

Systematization of Knowledge Papers

SoK: The Hitchhiker’s Guide to Efficient, End-to-End, and Tight DP Auditing
Meenatchi Sundaram Muthu Selva Annamalai (University College London), Borja Balle, Jamie Hayes, Georgios Kaissis (Google DeepMind) and Emiliano De Cristofaro (UC Riverside)
SoK: On the Survivability of Backdoor Attacks on Unconstrained Face Recognition Systems
Quentin Le Roux (Thales Group, Inria), Yannick Teglia (Thales Group), Teddy Furon (Inria), Philippe Loubet Moundi and Eric Bourbao (Thales Group)
SoK: Data Minimization in Machine Learning
Robin Staab, Nikola Jovanović (ETH Zurich), Kimberly Mai (University College London), Prakhar Ganesh (McGill University / Mila), Martin Vechev (ETH Zurich), Ferdinando Fioretto (University of Virginia) and Matthew Jagielski (Anthropic)
SoK: Decentralized AI (DeAI)
Elizabeth Lui (FLock.io), Rui Sun (Newcastle University & University of Manchester), Vatsal Shah (FLock.io), Xihan Xiong (Imperial College London), Jiahao Sun (FLock.io), Davide Crapis (Ethereum Foundation & PIN AI), William Knottenbelt (Imperial College London) and Zhipeng Wang (University of Manchester)
SoK: Privacy Risks and Mitigations in Retrieval-Augmented Generation Systems
Andreea-Elena Bodea, Stephen Meisenbacher, Alexandra Klymenko and Florian Matthes (Technical University of Munich)
SoK: Enhancing Cryptographic Collaborative Learning with Differential Privacy
Francesco Capano, Jonas Boehler (SAP SE) and Benjamin Weggenmann (Technische Hochschule Würzburg Schweinfurt)

Position Papers

Position: Research in Collaborative Learning Does Not Serve Cross-Silo Federated Learning in Practice
Kevin Kuo, Chhavi Yadav and Virginia Smith (Carnegie Mellon University)
Position: Gaussian DP for Reporting Differential Privacy Guarantees in Machine Learning
Juan Felipe Gomez (Harvard University), Bogdan Kulynych (Lausanne University Hospital), Georgios Kaissis, Borja Balle, Jamie Hayes (Google DeepMind), Flavio du Pin Calmon (Harvard University) and Antti Honkela (University of Helsinki)