A Complete List of All (arXiv) Adversarial Example Papers

by Nicholas Carlini 2019-06-15

It can be hard to stay up-to-date on the published papers in the field of adversarial examples, where we have seen massive growth in the number of papers written each year. I have been somewhat religiously keeping track of these papers for the last few years, and realized it may be helpful for others to release this list.

The only requirement I used for selecting papers for this list is that it is primarily a paper about adversarial examples, or extensively uses adversarial examples. Due to the sheer quantity of papers, I can't guarantee that I actually have found all of them.

But I did try.

I also may have included papers that don't match these criteria (and are about something different instead), or made inconsistent judgement calls as to whether or not any given paper is mainly an adversarial example paper. Send me an email if something is wrong and I'll correct it.

As a result, this list is completely un-filtered. Everything that mainly presents itself as an adversarial example paper is listed here; I pass no judgement of quality. For a curated list of papers that I think are excellent and worth reading, see the Adversarial Machine Learning Reading List.

One final note about the data. This list automatically updates with new papers, even before I get a chance to manually filter through them. I do this filtering roughly twice a week, and it's then that I'll remove the ones that aren't related to adversarial examples. As a result, there may be some false positives on the most recent few entries. The new un-verified entries will have a probability indicated that my simplistic (but reasonably well calibrated) bag-of-words classifier believes the given paper is actually about adversarial examples.

The full paper list appears below. I've also released a TXT file (and a TXT file with abstracts) and a JSON file with the same data. If you do anything interesting with this data I'd be happy to hear from you what it was.

Paper List

2024-04-18 Fortify the Guardian, Not the Treasure: Resilient Adversarial Detectors. (99%)Raz Lapid; Almog Dubin; Moshe Sipper Advancing the Robustness of Large Language Models through Self-Denoised Smoothing. (98%)Jiabao Ji; Bairu Hou; Zhen Zhang; Guanhua Zhang; Wenqi Fan; Qing Li; Yang Zhang; Gaowen Liu; Sijia Liu; Shiyu Chang Enhance Robustness of Language Models Against Variation Attack through Graph Integration. (33%)Zi Xiong; Lizhi Qing; Yangyang Kang; Jiawei Liu; Hongsong Li; Changlong Sun; Xiaozhong Liu; Wei Lu Uncovering Safety Risks in Open-source LLMs through Concept Activation Vector. (13%)Zhihao Xu; Ruixuan Huang; Xiting Wang; Fangzhao Wu; Jing Yao; Xing Xie Omniview-Tuning: Boosting Viewpoint Invariance of Vision-Language Pre-training Models. (2%)Shouwei Ruan; Yinpeng Dong; Hanqing Liu; Yao Huang; Hang Su; Xingxing Wei 2024-04-17 The Victim and The Beneficiary: Exploiting a Poisoned Model to Train a Clean Model on Poisoned Data. (83%)Zixuan Zhu; Rui Wang; Cong Zou; Lihua Jing GenFighter: A Generative and Evolutive Textual Attack Removal. (82%)Md Athikul Islam; Edoardo Serra; Sushil Jajodia Utilizing Adversarial Examples for Bias Mitigation and Accuracy Enhancement. (80%)Pushkar Shukla; Dhruv Srikanth; Lee Cohen; Matthew Turk Exploring DNN Robustness Against Adversarial Attacks Using Approximate Multipliers. (75%)Mohammad Javad Askarizadeh; Ebrahim Farahmand; Jorge Castro-Godinez; Ali Mahani; Laura Cabrera-Quiros; Carlos Salazar-Garcia Exploring the Transferability of Visual Prompting for Multimodal Large Language Models. (2%)Yichi Zhang; Yinpeng Dong; Siyuan Zhang; Tianzan Min; Hang Su; Jun Zhu Toward Understanding the Disagreement Problem in Neural Network Feature Attribution. (1%)Niklas Koenen; Marvin N. Wright Detector Collapse: Backdooring Object Detection to Catastrophic Overload or Blindness. (1%)Hangtao Zhang; Shengshan Hu; Yichen Wang; Leo Yu Zhang; Ziqi Zhou; Xianlong Wang; Yanjun Zhang; Chao Chen 2024-04-16 Efficiently Adversarial Examples Generation for Visual-Language Models under Targeted Transfer Scenarios using Diffusion Models. (99%)Qi Guo; Shanmin Pang; Xiaojun Jia; Qing Guo Adversarial Identity Injection for Semantic Face Image Synthesis. (38%)Giuseppe Tarollo; Tomaso Fontanini; Claudio Ferrari; Guido Borghi; Andrea Prati Robust Noisy Label Learning via Two-Stream Sample Distillation. (1%)Sihan Bai; Sanping Zhou; Zheng Qin; Le Wang; Nanning Zheng 2024-04-15 Black-box Adversarial Transferability: An Empirical Study in Cybersecurity Perspective. (99%)Khushnaseeb Roshan; Aasim Zafar Towards a Novel Perspective on Adversarial Examples Driven by Frequency. (99%)Zhun Zhang; Yi Zeng; Qihe Liu; Shijie Zhou Ti-Patch: Tiled Physical Adversarial Patch for no-reference video quality metrics. (83%)Victoria Leonenkova; Ekaterina Shumitskaya; Anastasia Antsiferova; Dmitriy Vatolin Improving Weakly-Supervised Object Localization Using Adversarial Erasing and Pseudo Label. (1%)Byeongkeun Kang; Sinhae Cha; Yeejin Lee Enhancing Code Vulnerability Detection via Vulnerability-Preserving Data Augmentation. (1%)Shangqing Liu; Wei Ma; Jian Wang; Xiaofei Xie; Ruitao Feng; Yang Liu Consistency and Uncertainty: Identifying Unreliable Responses From Black-Box Vision-Language Models for Selective Visual Question Answering. (1%)Zaid Khan; Yun Fu 2024-04-14 Watermark-embedded Adversarial Examples for Copyright Protection against Diffusion Models. (96%)Peifei Zhu; Tsubasa Takahashi; Hirokatsu Kataoka Counteracting Concept Drift by Learning with Future Malware Predictions. (96%)Branislav Bosansky; Lada Hospodkova; Michal Najman; Maria Rigaki; Elnaz Babayeva; Viliam Lisy Adversarial Robustness Limits via Scaling-Law and Human-Alignment Studies. (76%)Brian R. Bartoldson; James Diffenderfer; Konstantinos Parasyris; Bhavya Kailkhura FaceCat: Enhancing Face Recognition Security with a Unified Generative Model Framework. (22%)Jiawei Chen; Xiao Yang; Yinpeng Dong; Hang Su; Jianteng Peng; Zhaoxia Yin 2024-04-13 Stability and Generalization in Free Adversarial Training. (96%)Xiwei Cheng; Kexin Fu; Farzan Farnia Proof-of-Learning with Incentive Security. (2%)Zishuo Zhao; Zhixuan Fang; Xuechao Wang; Yuan Zhou 2024-04-12 PASA: Attack Agnostic Unsupervised Adversarial Detection using Prediction & Attribution Sensitivity Analysis. (99%)Dipkamal Bhusal; Md Tanvirul Alam; Monish K. Veerabhadran; Michael Clifford; Sara Rampazzi; Nidhi Rastogi Counterfactual Explanations for Face Forgery Detection via Adversarial Removal of Artifacts. (99%)Yang Li; Songlin Yang; Wei Wang; Ziwen He; Bo Peng; Jing Dong Struggle with Adversarial Defense? Try Diffusion. (99%)Yujie Li; Yanbin Wang; Haitao xu; Bin Liu; Jianguo Sun; Zhenhao Guo; Wenrui Ma Multimodal Attack Detection for Action Recognition Models. (83%)Furkan Mumcu; Yasin Yilmaz A Survey of Neural Network Robustness Assessment in Image Recognition. (83%)Jie Wang; Jun Ai; Minyan Lu; Haoran Su; Dan Yu; Yutao Zhang; Junda Zhu; Jingyu Liu Practical Region-level Attack against Segment Anything Models. (81%)Yifan Shen; Zhengyuan Li; Gang Wang FCert: Certifiably Robust Few-Shot Classification in the Era of Foundation Models. (69%)Yanting Wang; Wei Zou; Jinyuan Jia On the Robustness of Language Guidance for Low-Level Vision Tasks: Findings from Depth Estimation. (1%)Agneet Chatterjee; Tejas Gokhale; Chitta Baral; Yezhou Yang Empowering Malware Detection Efficiency within Processing-in-Memory Architecture. (1%)Sreenitha Kasarapu; Sathwika Bavikadi; Sai Manoj Pudukotai Dinakarrao 2024-04-11 Persistent Classification: A New Approach to Stability of Data and Adversarial Examples. (98%)Brian Bell; Michael Geyer; David Glickenstein; Keaton Hamm; Carlos Scheidegger; Amanda Fernandez; Juston Moore Eliminating Catastrophic Overfitting Via Abnormal Adversarial Examples Regularization. (98%)Runqi Lin; Chaojian Yu; Tongliang Liu Backdoor Contrastive Learning via Bi-level Trigger Optimization. (96%)Weiyu Sun; Xinyu Zhang; Hao Lu; Yingcong Chen; Ting Wang; Jinghui Chen; Lu Lin AmpleGCG: Learning a Universal and Transferable Generative Model of Adversarial Suffixes for Jailbreaking Both Open and Closed LLMs. (12%)Zeyi Liao; Huan Sun LeapFrog: The Rowhammer Instruction Skip Attack. (4%)Andrew Adiletta; Caner Tol; Berk Sunar Scaling (Down) CLIP: A Comprehensive Analysis of Data, Architecture, and Training Strategies. (1%)Zichao Li; Cihang Xie; Ekin Dogus Cubuk 2024-04-10 Logit Calibration and Feature Contrast for Robust Federated Learning on Non-IID Data. (99%)Yu Qiao; Chaoning Zhang; Apurba Adhikary; Choong Seon Hong Lost in Translation: Modern Neural Networks Still Struggle With Small Realistic Image Transformations. (82%)Ofir Shifman; Yair Weiss Adversarial purification for no-reference image-quality metrics: applicability study and new methods. (26%)Aleksandr Gushchin; Anna Chistyakova; Vladislav Minashkin; Anastasia Antsiferova; Dmitriy Vatolin Simpler becomes Harder: Do LLMs Exhibit a Coherent Behavior on Simplified Corpora? (2%)Miriam Anschütz; Edoardo Mosca; Georg Groh TrajPRed: Trajectory Prediction with Region-based Relation Learning. (1%)Chen Zhou; Ghassan AlRegib; Armin Parchami; Kunjan Singh 2024-04-09 Towards Building a Robust Toxicity Predictor. (99%)Dmitriy Bespalov; Sourav Bhabesh; Yi Xiang; Liutong Zhou; Yanjun Qi On adversarial training and the 1 Nearest Neighbor classifier. (99%)Amir Hagai; Yair Weiss LRR: Language-Driven Resamplable Continuous Representation against Adversarial Tracking Attacks. (80%)Jianlang Chen; Xuhong Ren; Qing Guo; Felix Juefei-Xu; Di Lin; Wei Feng; Lei Ma; Jianjun Zhao Towards Robust Domain Generation Algorithm Classification. (80%)Arthur Drichel; Marc Meyer; Ulrike Meyer SafeGen: Mitigating Unsafe Content Generation in Text-to-Image Models. (38%)Xinfeng Li; Yuchen Yang; Jiangyi Deng; Chen Yan; Yanjiao Chen; Xiaoyu Ji; Wenyuan Xu Sandwich attack: Multi-language Mixture Adaptive Attack on LLMs. (31%)Bibek Upadhayay; Vahid Behzadan Aggressive or Imperceptible, or Both: Network Pruning Assisted Hybrid Byzantines in Federated Learning. (26%)Emre Ozfatura; Kerem Ozfatura; Alptekin Kupcu; Deniz Gunduz How to Craft Backdoors with Unlabeled Data Alone? (1%)Yifei Wang; Wenhan Ma; Yisen Wang 2024-04-08 Certified PEFTSmoothing: Parameter-Efficient Fine-Tuning with Randomized Smoothing. (99%)Chengyan Fu; Wenjie Wang David and Goliath: An Empirical Evaluation of Attacks and Defenses for QNNs at the Deep Edge. (99%)Miguel Costa; Sandro Pinto BruSLeAttack: A Query-Efficient Score-Based Black-Box Sparse Adversarial Attack. (99%)Viet Quoc Vo; Ehsan Abbasnejad; Damith C. Ranasinghe Case Study: Neural Network Malware Detection Verification for Feature and Image Datasets. (98%)Preston K. Robinette; Diego Manzanas Lopez; Serena Serbinowska; Kevin Leach; Taylor T. Johnson Out-of-Distribution Data: An Acquaintance of Adversarial Examples -- A Survey. (98%)Naveen Karunanayake; Ravin Gunawardena; Suranga Seneviratne; Sanjay Chawla Quantum Adversarial Learning for Kernel Methods. (75%)Giuseppe Montalbano; Leonardo Banchi Investigating the Impact of Quantization on Adversarial Robustness. (50%)Qun Li; Yuan Meng; Chen Tang; Jiacheng Jiang; Zhi Wang SphereHead: Stable 3D Full-head Synthesis with Spherical Tri-plane Representation. (1%)Heyuan Li; Ce Chen; Tianhao Shi; Yuda Qiu; Sizhe An; Guanying Chen; Xiaoguang Han 2024-04-07 Semantic Stealth: Adversarial Text Attacks on NLP Using Several Methods. (99%)Roopkatha Dey; Aivy Debnath; Sayak Kumar Dutta; Kaustav Ghosh; Arijit Mitra; Arghya Roy Chowdhury; Jaydip Sen Enabling Privacy-Preserving Cyber Threat Detection with Federated Learning. (15%)Yu Bi; Yekai Li; Xuan Feng; Xianghang Mi How much reliable is ChatGPT's prediction on Information Extraction under Input Perturbations? (5%)Ishani Mondal; Abhilasha Sancheti SemEval-2024 Task 2: Safe Biomedical Natural Language Inference for Clinical Trials. (1%)Mael Jullien; Marco Valentino; André Freitas 2024-04-06 CANEDERLI: On The Impact of Adversarial Training and Transferability on CAN Intrusion Detection Systems. (86%)Francesco Marchiori; Mauro Conti Learning Minimal NAP Specifications for Neural Network Verification. (78%)Chuqin Geng; Zhaoyue Wang; Haolin Ye; Saifei Liao; Xujie Si Goal-guided Generative Prompt Injection Attack on Large Language Models. (67%)Chong Zhang; Mingyu Jin; Qinkai Yu; Chengzhi Liu; Haochen Xue; Xiaobo Jin Data Poisoning Attacks on Off-Policy Policy Evaluation Methods. (67%)Elita Lobo; Harvineet Singh; Marek Petrik; Cynthia Rudin; Himabindu Lakkaraju Structured Gradient-based Interpretations via Norm-Regularized Adversarial Training. (61%)Shizhan Gong; Qi Dou; Farzan Farnia Exploiting Sequence Number Leakage: TCP Hijacking in NAT-Enabled Wi-Fi Networks. (3%)Yuxiang Yang; Xuewei Feng; Qi Li; Kun Sun; Ziqiang Wang; Ke Xu 2024-04-05 Evaluating Adversarial Robustness: A Comparison Of FGSM, Carlini-Wagner Attacks, And The Role of Distillation as Defense Mechanism. (99%)Trilokesh Ranjan Sarkar; Nilanjan Das; Pralay Sankar Maitra; Bijoy Some; Ritwik Saha; Orijita Adhikary; Bishal Bose; Jaydip Sen Reliable Feature Selection for Adversarially Robust Cyber-Attack Detection. (98%)João Vitorino; Miguel Silva; Eva Maia; Isabel Praça Compositional Estimation of Lipschitz Constants for Deep Neural Networks. (13%)Yuezhu Xu; S. Sivaranjani Precision Guided Approach to Mitigate Data Poisoning Attacks in Federated Learning. (12%)K Naveen Kumar; C Krishna Mohan; Aravind Machiry 2024-04-04 Meta Invariance Defense Towards Generalizable Robustness to Unknown Adversarial Attacks. (99%)Lei Zhang; Yuhang Zhou; Yi Yang; Xinbo Gao FACTUAL: A Novel Framework for Contrastive Learning Based Robust SAR Image Classification. (98%)Xu Wang; Tian Ye; Rajgopal Kannan; Viktor Prasanna Learn What You Want to Unlearn: Unlearning Inversion Attacks against Machine Unlearning. (16%)Hongsheng Hu; Shuo Wang; Tian Dong; Minhui Xue Red Teaming GPT-4V: Are GPT-4V Safe Against Uni/Multi-Modal Jailbreak Attacks? (2%)Shuo Chen; Zhen Han; Bailan He; Zifeng Ding; Wenqian Yu; Philip Torr; Volker Tresp; Jindong Gu Knowledge Distillation-Based Model Extraction Attack using Private Counterfactual Explanations. (2%)Fatima Ezzeddine; Omran Ayoub; Silvia Giordano 2024-04-03 Adversarial Attacks and Dimensionality in Text Classifiers. (99%)Nandish Chattopadhyay; Atreya Goswami; Anupam Chattopadhyay Unsegment Anything by Simulating Deformation. (97%)Jiahao Lu; Xingyi Yang; Xinchao Wang "Are Adversarial Phishing Webpages a Threat in Reality?" Understanding the Users' Perception of Adversarial Webpages. (81%)Ying Yuan; Qingying Hao; Giovanni Apruzzese; Mauro Conti; Gang Wang JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak Attacks. (75%)Weidi Luo; Siyuan Ma; Xiaogeng Liu; Xiaoyu Guo; Chaowei Xiao Learn to Disguise: Avoid Refusal Responses in LLM's Defense via a Multi-agent Attacker-Disguiser Game. (11%)Qianqiao Xu; Zhiliang Tian; Hongyan Wu; Zhen Huang; Yiping Song; Feng Liu; Dongsheng Li A Unified Membership Inference Method for Visual Self-supervised Encoder via Part-aware Capability. (9%)Jie Zhu; Jirong Zha; Ding Li; Leye Wang Steganographic Passport: An Owner and User Verifiable Credential for Deep Model IP Protection Without Retraining. (1%)Qi Cui; Ruohan Meng; Chaohui Xu; Chip-Hong Chang 2024-04-02 Humanizing Machine-Generated Content: Evading AI-Text Detection through Adversarial Attack. (99%)Ying Zhou; Ben He; Le Sun ADVREPAIR:Provable Repair of Adversarial Attack. (99%)Zhiming Chi; Jianan Ma; Pengfei Yang; Cheng-Chao Huang; Renjue Li; Xiaowei Huang; Lijun Zhang Jailbreaking Prompt Attack: A Controllable Adversarial Attack against Diffusion Models. (97%)Jiachen Ma; Anda Cao; Zhiqing Xiao; Jie Zhang; Chao Ye; Junbo Zhao One Noise to Rule Them All: Multi-View Adversarial Attacks with Universal Perturbation. (92%)Mehmet Ergezer; Phat Duong; Christian Green; Tommy Nguyen; Abdurrahman Zeybey Defense without Forgetting: Continual Adversarial Defense with Anisotropic & Isotropic Pseudo Replay. (88%)Yuhang Zhou; Zhongyun Hua Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks. (82%)Maksym Andriushchenko; Francesco Croce; Nicolas Flammarion READ: Improving Relation Extraction from an ADversarial Perspective. (81%)Dawei Li; William Hogan; Jingbo Shang Two Heads are Better than One: Nested PoE for Robust Defense Against Multi-Backdoors. (64%)Victoria Graf; Qin Liu; Muhao Chen Red-Teaming Segment Anything Model. (45%)Krzysztof Jankowski; Bartlomiej Sobieski; Mateusz Kwiatkowski; Jakub Szulc; Michal Janik; Hubert Baniecki; Przemyslaw Biecek Towards Robust 3D Pose Transfer with Adversarial Learning. (31%)Haoyu Chen; Hao Tang; Ehsan Adeli; Guoying Zhao Designing a Photonic Physically Unclonable Function Having Resilience to Machine Learning Attacks. (12%)Elena R. Henderson; Jessie M. Henderson; Hiva Shahoei; William V. Oxford; Eric C. Larson; Duncan L. MacFarlane; Mitchell A. Thornton Exploring Backdoor Vulnerabilities of Chat Models. (2%)Yunzhuo Hao; Wenkai Yang; Yankai Lin CAPE: CAM as a Probabilistic Ensemble for Enhanced DNN Interpretation. (1%)Townim Faisal Chowdhury; Kewen Liao; Vu Minh Hieu Phan; Minh-Son To; Yutong Xie; Kevin Hung; David Ross; Anton van den Hengel; Johan W. Verjans; Zhibin Liao 2024-04-01 The Double-Edged Sword of Input Perturbations to Robust Accurate Fairness. (99%)Xuran Li; Peng Wu; Yanting Chen; Xingjun Ma; Zhen Zhang; Kaixiang Dong Multi-granular Adversarial Attacks against Black-box Neural Ranking Models. (99%)Yu-An Liu; Ruqing Zhang; Jiafeng Guo; Rijke Maarten de; Yixing Fan; Xueqi Cheng BadPart: Unified Black-box Adversarial Patch Attacks against Pixel-wise Regression Tasks. (93%)Zhiyuan Cheng; Zhaoyi Liu; Tengda Guo; Shiwei Feng; Dongfang Liu; Mingjie Tang; Xiangyu Zhang Poisoning Decentralized Collaborative Recommender System and Its Countermeasures. (33%)Ruiqi Zheng; Liang Qu; Tong Chen; Kai Zheng; Yuhui Shi; Hongzhi Yin Can Biases in ImageNet Models Explain Generalization? (10%)Paul Gavrikov; Janis Keuper UFID: A Unified Framework for Input-level Backdoor Detection on Diffusion Models. (10%)Zihan Guan; Mengxuan Hu; Sheng Li; Anil Vullikanti Privacy Backdoors: Enhancing Membership Inference through Poisoning Pre-trained Models. (2%)Yuxin Wen; Leo Marchyok; Sanghyun Hong; Jonas Geiping; Tom Goldstein; Nicholas Carlini An incremental hybrid adaptive network-based IDS in Software Defined Networks to detect stealth attacks. (1%)Abdullah H Alqahtani 2024-03-31 PID Control-Based Self-Healing to Improve the Robustness of Large Language Models. (75%)Zhuotong Chen; Zihu Wang; Yifan Yang; Qianxiao Li; Zheng Zhang Machine Learning Robustness: A Primer. (62%)Houssem Ben Braiek; Foutse Khomh 2024-03-30 STBA: Towards Evaluating the Robustness of DNNs for Query-Limited Black-box Scenario. (99%)Renyang Liu; Kwok-Yan Lam; Wei Zhou; Sixing Wu; Jun Zhao; Dongting Hu; Mingming Gong Embodied Active Defense: Leveraging Recurrent Feedback to Counter Adversarial Patches. (98%)Lingxuan Wu; Xiao Yang; Yinpeng Dong; Liuwei Xie; Hang Su; Jun Zhu Shortcuts Arising from Contrast: Effective and Covert Clean-Label Attacks in Prompt-Based Learning. (5%)Xiaopeng Xie; Ming Yan; Xiwen Zhou; Chenlong Zhao; Suli Wang; Yong Zhang; Joey Tianyi Zhou 2024-03-29 On Inherent Adversarial Robustness of Active Vision Systems. (99%)Amitangshu Mukherjee; Timur Ibrayev; Kaushik Roy Benchmarking the Robustness of Temporal Action Detection Models Against Temporal Corruptions. (68%)Runhao Zeng; Xiaoyong Chen; Jiaming Liang; Huisi Wu; Guangzhong Cao; Yong Guo Deepfake Sentry: Harnessing Ensemble Intelligence for Resilient Detection and Generalisation. (8%)Liviu-Daniel University "Politehnica" of Bucharest, Romania Ştefan; Dan-Cristian University "Politehnica" of Bucharest, Romania Stanciu; Mihai University "Politehnica" of Bucharest, Romania Dogariu; Mihai Gabriel University "Politehnica" of Bucharest, Romania Constantin; Andrei Cosmin University "Politehnica" of Bucharest, Romania Jitaru; Bogdan University "Politehnica" of Bucharest, Romania Ionescu The Impact of Prompts on Zero-Shot Detection of AI-Generated Text. (2%)Kaito Taguchi; Yujie Gu; Kouichi Sakurai GDA: Generalized Diffusion for Robust Test-time Adaptation. (1%)Yun-Yun Tsai; Fu-Chen Chen; Albert Y. C. Chen; Junfeng Yang; Che-Chun Su; Min Sun; Cheng-Hao Kuo Efficient Data-Free Model Stealing with Label Diversity. (1%)Yiyong Liu; Rui Wen; Michael Backes; Yang Zhang Cross-Lingual Transfer Robustness to Lower-Resource Languages on Adversarial Datasets. (1%)Shadi Manafi; Nikhil Krishnaswamy 2024-03-28 Towards Understanding Dual BN In Hybrid Adversarial Training. (82%)Chenshuang Zhang; Chaoning Zhang; Kang Zhang; Axi Niu; Junmo Kim; In So Kweon Improving Adversarial Data Collection by Supporting Annotators: Lessons from GAHD, a German Hate Speech Dataset. (82%)Janis Goldzycher; Paul Röttger; Gerold Schneider On the Robustness of LDP Protocols for Numerical Attributes under Data Poisoning Attacks. (41%)Xiaoguang Li; Zitao Li; Ninghui Li; Wenhai Sun MedBN: Robust Test-Time Adaptation against Malicious Test Samples. (10%)Hyejin Park; Jeongyeon Hwang; Sunung Mun; Sangdon Park; Jungseul Ok A Backdoor Approach with Inverted Labels Using Dirty Label-Flipping Attacks. (1%)Orson Mengara Imperceptible Protection against Style Imitation from Diffusion Models. (1%)Namhyuk Ahn; Wonhyuk Ahn; KiYoon Yoo; Daesik Kim; Seung-Hun Nam 2024-03-27 Uncertainty-Aware SAR ATR: Defending Against Adversarial Attacks via Bayesian Neural Networks. (99%)Tian Ye; Rajgopal Kannan; Viktor Prasanna; Carl Busart CosalPure: Learning Concept from Group Images for Robust Co-Saliency Detection. (99%)Jiayi Zhu; Qing Guo; Felix Juefei-Xu; Yihao Huang; Yang Liu; Geguang Pu MMCert: Provable Defense against Adversarial Attacks to Multi-modal Models. (98%)Yanting Wang; Hongye Fu; Wei Zou; Jinyuan Jia Bayesian Learned Models Can Detect Adversarial Malware For Free. (97%)Bao Gia Doan; Dang Quang Nguyen; Paul Montague; Tamas Abraham; Vel Olivier De; Seyit Camtepe; Salil S. Kanhere; Ehsan Abbasnejad; Damith C. Ranasinghe MisGUIDE : Defense Against Data-Free Deep Learning Model Extraction. (95%)Mahendra Gurve; Sankar Behera; Satyadev Ahlawat; Yamuna Prasad Towards Sustainable SecureML: Quantifying Carbon Footprint of Adversarial Machine Learning. (83%)Syed Mhamudul Hasan; Abdur R. Shahid; Ahmed Imteaj Deep Learning for Robust and Explainable Models in Computer Vision. (82%)Mohammadreza Amirian SemRoDe: Macro Adversarial Training to Learn Representations That are Robust to Word-Level Attacks. (81%)Brian Formento; Wenjie Feng; Chuan Sheng Foo; Luu Anh Tuan; See-Kiong Ng JailbreakBench: An Open Robustness Benchmark for Jailbreaking Large Language Models. (45%)Patrick Chao; Edoardo Debenedetti; Alexander Robey; Maksym Andriushchenko; Francesco Croce; Vikash Sehwag; Edgar Dobriban; Nicolas Flammarion; George J. Pappas; Florian Tramer; Hamed Hassani; Eric Wong Vulnerability Detection with Code Language Models: How Far Are We? (26%)Yangruibo Ding; Yanjun Fu; Omniyyah Ibrahim; Chawin Sitawarin; Xinyun Chen; Basel Alomair; David Wagner; Baishakhi Ray; Yizheng Chen Spikewhisper: Temporal Spike Backdoor Attacks on Federated Neuromorphic Learning over Low-power Devices. (15%)Hanqing Fu; Gaolei Li; Jun Wu; Jianhua Li; Xi Lin; Kai Zhou; Yuchen Liu Robustness and Visual Explanation for Black Box Image, Video, and ECG Signal Classification with Reinforcement Learning. (15%)Soumyendu Sarkar; Ashwin Ramesh Babu; Sajad Mousavi; Vineet Gundecha; Avisek Naug; Sahand Ghorbanpour The Impact of Uniform Inputs on Activation Sparsity and Energy-Latency Attacks in Computer Vision. (11%)Andreas Müller; Erwin Quiring Fact Checking Beyond Training Set. (1%)Payam Karisani; Heng Ji BAM: Box Abstraction Monitors for Real-time OoD Detection in Object Detection. (1%)Changshun Wu; Weicheng He; Chih-Hong Cheng; Xiaowei Huang; Saddek Bensalem 2024-03-26 DataCook: Crafting Anti-Adversarial Examples for Healthcare Data Copyright Protection. (92%)Sihan Shang; Jiancheng Yang; Zhenglong Sun; Pascal Fua FaultGuard: A Generative Approach to Resilient Fault Prediction in Smart Electrical Grids. (78%)Emad Efatinasab; Francesco Marchiori; Alessandro Brighente; Mirco Rampazzo; Mauro Conti Boosting Adversarial Training via Fisher-Rao Norm-based Regularization. (69%)Xiangyu Yin; Wenjie Ruan Targeted Visualization of the Backbone of Encoder LLMs. (9%)Isaac Roberts; Alexander Schulz; Luca Hermes; Barbara Hammer Optimization-based Prompt Injection Attack to LLM-as-a-Judge. (4%)Jiawen Shi; Zenghui Yuan; Yinuo Liu; Yue Huang; Pan Zhou; Lichao Sun; Neil Zhenqiang Gong Leak and Learn: An Attacker's Cookbook to Train Using Leaked Data from Federated Learning. (1%)Joshua C. Zhao; Ahaan Dabholkar; Atul Sharma; Saurabh Bagchi Exploring LLMs as a Source of Targeted Synthetic Textual Data to Minimize High Confidence Misclassifications. (1%)Philip Lippmann; Matthijs Spaan; Jie Yang 2024-03-25 $\textit{LinkPrompt}$: Natural and Universal Adversarial Attacks on Prompt-based Language Models. (99%)Yue Xu; Wenjie Wang Physical 3D Adversarial Attacks against Monocular Depth Estimation in Autonomous Driving. (98%)Junhao Zheng; Chenhao Lin; Jiahao Sun; Zhengyu Zhao; Qian Li; Chao Shen The Anatomy of Adversarial Attacks: Concept-based XAI Dissection. (87%)Georgii Mikriukov; Gesina Schwalbe; Franz Motzkus; Korinna Bade DeepKnowledge: Generalisation-Driven Deep Learning Testing. (82%)Sondess Missaoui; Simos Gerasimou; Nikolaos Matragkas Revealing Vulnerabilities of Neural Networks in Parameter Learning and Defense Against Explanation-Aware Backdoors. (70%)Md Abdul Kadir; GowthamKrishna Addluri; Daniel Sonntag LOTUS: Evasive and Resilient Backdoor Attacks through Sub-Partitioning. (69%)Siyuan Cheng; Guanhong Tao; Yingqi Liu; Guangyu Shen; Shengwei An; Shiwei Feng; Xiangzhe Xu; Kaiyuan Zhang; Shiqing Ma; Xiangyu Zhang Model-less Is the Best Model: Generating Pure Code Implementations to Replace On-Device DL Models. (1%)Mingyi Zhou; Xiang Gao; Pei Liu; John Grundy; Chunyang Chen; Xiao Chen; Li Li 2024-03-24 Subspace Defense: Discarding Adversarial Perturbations by Learning a Subspace for Clean Signals. (99%)Rui Zheng; Yuhao Zhou; Zhiheng Xi; Tao Gui; Qi Zhang; Xuanjing Huang Ensemble Adversarial Defense via Integration of Multiple Dispersed Low Curvature Models. (98%)Kaikang Zhao; Xi Chen; Wei Huang; Liuxin Ding; Xianglong Kong; Fan Zhang Robust Diffusion Models for Adversarial Purification. (73%)Guang Lin; Zerui Tao; Jianhai Zhang; Toshihisa Tanaka; Qibin Zhao Unlearning Backdoor Threats: Enhancing Backdoor Defense in Multimodal Contrastive Learning via Local Token Unlearning. (5%)Siyuan Liang; Kuanrong Liu; Jiajun Gong; Jiawei Liang; Yuan Xun; Ee-Chien Chang; Xiaochun Cao Rumor Detection with a novel graph neural network approach. (4%)Tianrui Liu; Qi Cai; Changxin Xu; Bo Hong; Fanghao Ni; Yuxin Qiao; Tsungwei Yang Generating Potent Poisons and Backdoors from Scratch with Guided Diffusion. (2%)Hossein Souri; Arpit Bansal; Hamid Kazemi; Liam Fowl; Aniruddha Saha; Jonas Geiping; Andrew Gordon Wilson; Rama Chellappa; Tom Goldstein; Micah Goldblum A General and Efficient Federated Split Learning with Pre-trained Image Transformers for Heterogeneous Data. (1%)Yifan Shi; Yuhui Zhang; Ziyue Huang; Xiaofeng Yang; Li Shen; Wei Chen; Xueqian Wang 2024-03-23 An Embarrassingly Simple Defense Against Backdoor Attacks On SSL. (70%)Aryan Satpathy; Nilaksh; Dhruva Rajwade Adversarial Defense Teacher for Cross-Domain Object Detection under Poor Visibility Conditions. (64%)Kaiwen Wang; Yinzhe Shen; Martin Lauer 2024-03-22 Robust optimization for adversarial learning with finite sample complexity guarantees. (96%)André Bertolace; Konstatinos Gatsis; Kostas Margellos A Transfer Attack to Image Watermarks. (95%)Yuepeng Hu; Zhengyuan Jiang; Moyang Guo; Neil Gong From Hardware Fingerprint to Access Token: Enhancing the Authentication on IoT Devices. (26%)Yue Xiao; Yi He; Xiaoli Zhang; Qian Wang; Renjie Xie; Kun Sun; Ke Xu; Qi Li Clean-image Backdoor Attacks. (12%)Dazhong Rong; Guoyao Yu; Shuheng Shen; Xinyi Fu; Peng Qian; Jianhai Chen; Qinming He; Xing Fu; Weiqiang Wang Forward Learning for Gradient-based Black-box Saliency Map Generation. (1%)Zeliang Zhang; Mingqian Feng; Jinyang Jiang; Rongyi Zhu; Yijie Peng; Chenliang Xu 2024-03-21 Diffusion Attack: Leveraging Stable Diffusion for Naturalistic Image Attacking. (99%)Qianyu Guo; Jiaming Fu; Yawen Lu; Dongming Gan Few-Shot Adversarial Prompt Learning on Vision-Language Models. (98%)Yiwei Zhou; Xiaobo Xia; Zhiwei Lin; Bo Han; Tongliang Liu Reversible Jump Attack to Textual Classifiers with Modification Reduction. (98%)Mingze Ni; Zhensu Sun; Wei Liu Improving Robustness to Model Inversion Attacks via Sparse Coding Architectures. (82%)Sayanton V. Dibbo; Adam Breuer; Juston Moore; Michael Teti Adversary-Robust Graph-Based Learning of WSIs. (45%)Saba Heidari Gheshlaghi; Milan Aryal; Nasim Yahyasoltani; Masoud Ganji Safeguarding Medical Image Segmentation Datasets against Unauthorized Training via Contour- and Texture-Aware Perturbations. (4%)Xun Lin; Yi Yu; Song Xia; Jue Jiang; Haoran Wang; Zitong Yu; Yizhong Liu; Ying Fu; Shuai Wang; Wenzhong Tang; Alex Kot 2024-03-20 FMM-Attack: A Flow-based Multi-modal Adversarial Attack on Video-based LLMs. (97%)Jinmin Li; Kuofeng Gao; Yang Bai; Jingyun Zhang; Shu-tao Xia; Yisen Wang DD-RobustBench: An Adversarial Robustness Benchmark for Dataset Distillation. (96%)Yifan Wu; Jiawei Du; Ping Liu; Yuewei Lin; Wenqing Cheng; Wei Xu Capsule Neural Networks as Noise Stabilizer for Time Series Data. (93%)Soyeon Kim; Jihyeon Seong; Hyunkyung Han; Jaesik Choi Adversarial Attacks and Defenses in Automated Control Systems: A Comprehensive Benchmark. (70%)Vitaliy Pozdnyakov; Aleksandr Kovalenko; Ilya Makarov; Mikhail Drobyshevskiy; Kirill Lukyanov Certified Human Trajectory Prediction. (61%)Mohammadhossein Bahari; Saeed Saadatnejad; Amirhossein Asgari Farsangi; Seyed-Mohsen Moosavi-Dezfooli; Alexandre Alahi Have You Poisoned My Data? Defending Neural Networks against Data Poisoning. (54%)Gaspari Fabio De; Dorjan Hitaj; Luigi V. Mancini Defending Against Indirect Prompt Injection Attacks With Spotlighting. (31%)Keegan Hines; Gary Lopez; Matthew Hall; Federico Zarfati; Yonatan Zunger; Emre Kiciman Don't be a Fool: Pooling Strategies in Offensive Language Detection from User-Intended Adversarial Attacks. (11%)Seunguk Yu; Juhwan Choi; Youngbin Kim BadEdit: Backdooring large language models by model editing. (1%)Yanzhou Li; Tianlin Li; Kangjie Chen; Jian Zhang; Shangqing Liu; Wenhan Wang; Tianwei Zhang; Yang Liu Teacher-Student Training for Debiasing: General Permutation Debiasing for Large Language Models. (1%)Adian Liusie; Yassir Fathullah; Mark J. F. Gales Threats, Attacks, and Defenses in Machine Unlearning: A Survey. (1%)Ziyao Liu; Huanyi Ye; Chen Chen; Kwok-Yan Lam 2024-03-19 As Firm As Their Foundations: Can open-sourced foundation models be used to create adversarial examples for downstream tasks? (99%)Anjun Hu; Jindong Gu; Francesco Pinto; Konstantinos Kamnitsas; Philip Torr Boosting Transferability in Vision-Language Attacks via Diversification along the Intersection Region of Adversarial Trajectory. (99%)Sensen Gao; Xiaojun Jia; Xuhong Ren; Ivor Tsang; Qing Guo ADAPT to Robustify Prompt Tuning Vision Transformers. (98%)Masih Eskandar; Tooba Imtiaz; Zifeng Wang; Jennifer Dy RigorLLM: Resilient Guardrails for Large Language Models against Undesired Content. (8%)Zhuowen Yuan; Zidi Xiong; Yi Zeng; Ning Yu; Ruoxi Jia; Dawn Song; Bo Li Robust NAS under adversarial training: benchmark, theory, and beyond. (2%)Yongtao Wu; Fanghui Liu; Carl-Johann Simon-Gabriel; Grigorios G Chrysos; Volkan Cevher Discover and Mitigate Multiple Biased Subgroups in Image Classifiers. (1%)Zeliang Zhang; Mingqian Feng; Zhiheng Li; Chenliang Xu 2024-03-18 Diffusion Denoising as a Certified Defense against Clean-label Poisoning. (99%)Sanghyun Hong; Nicholas Carlini; Alexey Kurakin SSCAE -- Semantic, Syntactic, and Context-aware natural language Adversarial Examples generator. (99%)Javad Rafiei Asl; Mohammad H. Rafiei; Manar Alohaly; Daniel Takabi LocalStyleFool: Regional Video Style Transfer Attack Using Segment Anything Model. (99%)Yuxin Cao; Jinghao Li; Xi Xiao; Derui Wang; Minhui Xue; Hao Ge; Wei Liu; Guangwu Hu Invisible Backdoor Attack Through Singular Value Decomposition. (96%)Wenmin Chen; Xiaowei Xu Problem space structural adversarial attacks for Network Intrusion Detection Systems based on Graph Neural Networks. (88%)Andrea Venturi; Dario Stabili; Mirco Marchetti Impart: An Imperceptible and Effective Label-Specific Backdoor Attack. (83%)Jingke Zhao; Zan Wang; Yongwei Wang; Lanjun Wang SSAP: A Shape-Sensitive Adversarial Patch for Comprehensive Disruption of Monocular Depth Estimation in Autonomous Navigation Applications. (78%)Amira Guesmi; Muhammad Abdullah Hanif; Ihsen Alouani; Bassem Ouni; Muhammad Shafique Electioneering the Network: Dynamic Multi-Step Adversarial Attacks for Community Canvassing. (61%)Saurabh Sharma; Ambuj SIngh Advancing Time Series Classification with Multimodal Language Modeling. (1%)Mingyue Cheng; Yiheng Chen; Qi Liu; Zhiding Liu; Yucong Luo 2024-03-17 Defense Against Adversarial Attacks on No-Reference Image Quality Models with Gradient Norm Regularization. (99%)Yujia Liu; Chenxi Yang; Dingquan Li; Jianhao Ding; Tingting Jiang A Modified Word Saliency-Based Adversarial Attack on Text Classification Models. (99%)Hetvi Waghela; Sneha Rakshit; Jaydip Sen Robust Overfitting Does Matter: Test-Time Adversarial Purification With FGSM. (99%)Linyu Tang; Lei Zhang Forging the Forger: An Attempt to Improve Authorship Verification via Data Augmentation. (76%)Silvia Corbara; Alejandro Moreo RobustSentEmbed: Robust Sentence Embeddings Using Adversarial Self-Supervised Contrastive Learning. (50%)Javad Rafiei Asl; Prajwal Panzade; Eduardo Blanco; Daniel Takabi; Zhipeng Cai COLEP: Certifiably Robust Learning-Reasoning Conformal Prediction via Probabilistic Circuits. (22%)Mintong Kang; Nezihe Merve Gürel; Linyi Li; Bo Li A Dual-Tier Adaptive One-Class Classification IDS for Emerging Cyberthreats. (9%)Md. Ashraf Uddin; Sunil Aryal; Mohamed Reda Bouadjenek; Muna Al-Hawawreh; Md. Alamin Talukder Hierarchical Classification for Intrusion Detection System: Effective Design and Empirical Analysis. (2%)Md. Ashraf Uddin; Sunil Aryal; Mohamed Reda Bouadjenek; Muna Al-Hawawreh; Md. Alamin Talukder CBR - Boosting Adaptive Classification By Retrieval of Encrypted Network Traffic with Out-of-distribution. (1%)Amir Lukach; Ran Dubin; Amit Dvir; Chen Hajaj Pencil: Private and Extensible Collaborative Learning without the Non-Colluding Assumption. (1%)Xuanqi Liu; Zhuotao Liu; Qi Li; Ke Xu; Mingwei Xu 2024-03-16 Securely Fine-tuning Pre-trained Encoders Against Adversarial Examples. (98%)Ziqi Zhou; Minghui Li; Wei Liu; Shengshan Hu; Yechao Zhang; Wei Wan; Lulu Xue; Leo Yu Zhang; Dezhong Yang; Hai Jin Understanding Robustness of Visual State Space Models for Image Classification. (98%)Chengbin Du; Yanxi Li; Chang Xu Improving Adversarial Transferability of Visual-Language Pre-training Models through Collaborative Multimodal Interaction. (92%)Jiyuan Fu; Zhaoyu Chen; Kaixun Jiang; Haijing Guo; Jiafeng Wang; Shuyong Gao; Wenqiang Zhang Edge Private Graph Neural Networks with Singular Value Perturbation. (11%)Tingting Tang; Yue Niu; Salman Avestimehr; Murali Annavaram 2024-03-15 Benchmarking Adversarial Robustness of Image Shadow Removal with Shadow-adaptive Attacks. (99%)Chong Wang; Yi Yu; Lanqing Guo; Bihan Wen Towards Non-Adversarial Algorithmic Recourse. (99%)Tobias Leemann; Martin Pawelczyk; Bardh Prenkaj; Gjergji Kasneci Time-Frequency Jointed Imperceptible Adversarial Attack to Brainprint Recognition with Deep Learning Models. (99%)Hangjie Yi; Yuhang Ming; Dongjun Liu; Wanzeng Kong Introducing Adaptive Continuous Adversarial Training (ACAT) to Enhance ML Robustness. (87%)Mohamed elShehaby; Aditya Kotha; Ashraf Matrawy Revisiting Adversarial Training under Long-Tailed Distributions. (80%)Xinli Yue; Ningping Mou; Qian Wang; Lingchen Zhao Towards Adversarially Robust Dataset Distillation by Curvature Regularization. (54%)Eric Xue; Yijiang Li; Haoyang Liu; Yifan Shen; Haohan Wang Interactive Trimming against Evasive Online Data Manipulation Attacks: A Game-Theoretic Approach. (50%)Yue Fu; Qingqing Ye; Rong Du; Haibo Hu Securing Federated Learning with Control-Flow Attestation: A Novel Framework for Enhanced Integrity and Resilience against Adversarial Attacks. (12%)Zahir Alsulaimawi Benchmarking Zero-Shot Robustness of Multimodal Foundation Models: A Pilot Study. (11%)Chenguang Wang; Ruoxi Jia; Xin Liu; Dawn Song Not Just Change the Labels, Learn the Features: Watermarking Deep Neural Networks with Multi-View Data. (4%)Yuxuan Li; Sarthak Kumar Maharana; Yunhui Guo Backdoor Secrets Unveiled: Identifying Backdoor Data with Optimized Scaled Prediction Consistency. (3%)Soumyadeep Pal; Yuguang Yao; Ren Wang; Bingquan Shen; Sijia Liu Robust Influence-based Training Methods for Noisy Brain MRI. (1%)Minh-Hao Van; Alycia N. Carey; Xintao Wu 2024-03-14 An Image Is Worth 1000 Lies: Adversarial Transferability across Prompts on Vision-Language Models. (99%)Haochen Luo; Jindong Gu; Fengyuan Liu; Philip Torr Counter-Samples: A Stateless Strategy to Neutralize Black Box Adversarial Attacks. (99%)Roey Bokobza; Yisroel Mirsky Adversarial Fine-tuning of Compressed Neural Networks for Joint Improvement of Robustness and Efficiency. (98%)Hallgrimur Thorsteinsson; Valdemar J Henriksen; Tong Chen; Raghavendra Selvan Soften to Defend: Towards Adversarial Robustness via Self-Guided Label Refinement. (83%)Daiwei Yu; Zhuorong Li; Lina Wei; Canghong Jin; Yun Zhang; Sixian Chan Robust Subgraph Learning by Monitoring Early Training Representations. (80%)Sepideh Neshatfar; Salimeh Yasaei Sekeh LDPRecover: Recovering Frequencies from Poisoning Attacks against Local Differential Privacy. (76%)Xinyue Sun; Qingqing Ye; Haibo Hu; Jiawei Duan; Tianyu Wo; Jie Xu; Renyu Yang AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting. (74%)Yu Wang; Xiaogeng Liu; Yu Li; Muhao Chen; Chaowei Xiao Towards White Box Deep Learning. (15%)Maciej Satkiewicz Symbiotic Game and Foundation Models for Cyber Deception Operations in Strategic Cyber Warfare. (13%)Tao Li; Quanyan Zhu PreCurious: How Innocent Pre-Trained Language Models Turn into Privacy Traps. (8%)Ruixuan Liu; Tianhao Wang; Yang Cao; Li Xiong AVIBench: Towards Evaluating the Robustness of Large Vision-Language Model on Adversarial Visual-Instructions. (2%)Hao Zhang; Wenqi Shao; Hong Liu; Yongqiang Ma; Ping Luo; Yu Qiao; Kaipeng Zhang Medical Unlearnable Examples: Securing Medical Data from Unauthorized Traning via Sparsity-Aware Local Masking. (1%)Weixiang Sun; Yixin Liu; Zhiling Yan; Kaidi Xu; Lichao Sun 2024-03-13 Attack Deterministic Conditional Image Generative Models for Diverse and Controllable Generation. (92%)Tianyi Chu; Wei Xing; Jiafu Chen; Zhizhong Wang; Jiakai Sun; Lei Zhao; Haibo Chen; Huaizhong Lin Fast Inference of Removal-Based Node Influence. (54%)Weikai Li; Zhiping Xiao; Xiao Luo; Yizhou Sun Tastle: Distract Large Language Models for Automatic Jailbreak Attack. (31%)Zeguan Xiao; Yan Yang; Guanhua Chen; Yun Chen Adaptive Hybrid Masking Strategy for Privacy-Preserving Face Recognition Against Model Inversion Attack. (8%)Yuanqing Huang; Yinggui Wang; Jianshu Li; Le Yang; Kai Song; Lei Wang RAF-GI: Towards Robust, Accurate and Fast-Convergent Gradient Inversion Attack in Federated Learning. (2%)Can Liu; Jin Wang; Dongyang Yu Verifix: Post-Training Correction to Improve Label Noise Robustness with Verified Samples. (1%)Sangamesh Kodge; Deepak Ravikumar; Gobinda Saha; Kaushik Roy 2024-03-12 Versatile Defense Against Adversarial Attacks on Image Recognition. (99%)Haibo Zhang; Zhihua Yao; Kouichi Sakurai Towards Model Extraction Attacks in GAN-Based Image Translation via Domain Shift Mitigation. (61%)Di Mi; Yanjun Zhang; Leo Yu Zhang; Shengshan Hu; Qi Zhong; Haizhuan Yuan; Shirui Pan Backdoor Attack with Mode Mixture Latent Modification. (8%)Hongwei Zhang; Xiaoyin Xu; Dongsheng An; Xianfeng Gu; Min Zhang Duwak: Dual Watermarks in Large Language Models. (2%)Chaoyi Zhu; Jeroen Galjaard; Pin-Yu Chen; Lydia Y. Chen Towards a Framework for Deep Learning Certification in Safety-Critical Applications Using Inherently Safe Design and Run-Time Error Detection. (2%)Romeo Valentin Visual Privacy Auditing with Diffusion Models. (1%)Kristian Schwethelm; Johannes Kaiser; Moritz Knolle; Daniel Rueckert; Georgios Kaissis; Alexander Ziller 2024-03-11 Intra-Section Code Cave Injection for Adversarial Evasion Attacks on Windows PE Malware File. (99%)Kshitiz Aryal; Maanak Gupta; Mahmoud Abdelsalam; Moustafa Saleh epsilon-Mesh Attack: A Surface-based Adversarial Point Cloud Attack for Facial Expression Recognition. (99%)Batuhan Cengiz; Mert Gulsen; Yusuf H. Sahin; Gozde Unal PeerAiD: Improving Adversarial Distillation from a Specialized Peer Tutor. (98%)Jaewon Jung; Hongsun Jang; Jaeyong Song; Jinho Lee Dynamic Perturbation-Adaptive Adversarial Training on Medical Image Classification. (97%)Shuai Li; Xiaoguang Ma; Shancheng Jiang; Lu Meng Disentangling Policy from Offline Task Representation Learning via Adversarial Data Augmentation. (96%)Chengxing Jia; Fuxiang Zhang; Yi-Chen Li; Chen-Xiao Gao; Xu-Hui Liu; Lei Yuan; Zongzhang Zhang; Yang Yu PCLD: Point Cloud Layerwise Diffusion for Adversarial Purification. (86%)Mert Gulsen; Batuhan Cengiz; Yusuf H. Sahin; Gozde Unal Overcoming the Paradox of Certified Training with Gaussian Smoothing. (81%)Stefan Balauca; Mark Niklas Müller; Yuhao Mao; Maximilian Baader; Marc Fischer; Martin Vechev Real is not True: Backdoor Attacks Against Deepfake Detection. (78%)Hong Sun; Ziqiang Li; Lei Liu; Bin Li Improving deep learning with prior knowledge and cognitive models: A survey on enhancing explainability, adversarial robustness and zero-shot learning. (61%)Fuseinin Mumuni; Alhassan Mumuni Stealing Part of a Production Language Model. (33%)Nicholas Carlini; Daniel Paleka; Krishnamurthy Dj Dvijotham; Thomas Steinke; Jonathan Hayase; A. Feder Cooper; Katherine Lee; Matthew Jagielski; Milad Nasr; Arthur Conmy; Eric Wallace; David Rolnick; Florian Tramèr AS-FIBA: Adaptive Selective Frequency-Injection for Backdoor Attack on Deep Face Restoration. (9%)Zhenbo Song; Wenhao Gao; Kaihao Zhang; Wenhan Luo; Zhaoxin Fan; Jianfeng Lu A novel interface for adversarial trivia question-writing. (3%)Jason Liu Towards the Uncharted: Density-Descending Feature Perturbation for Semi-supervised Semantic Segmentation. (2%)Xiaoyang Wang; Huihui Bai; Limin Yu; Yao Zhao; Jimin Xiao Learning with Noisy Foundation Models. (1%)Hao Chen; Jindong Wang; Zihan Wang; Ran Tao; Hongxin Wei; Xing Xie; Masashi Sugiyama; Bhiksha Raj DNNShield: Embedding Identifiers for Deep Neural Network Ownership Verification. (1%)Jasper Stang; Torsten Krauß; Alexandra Dmitrienko 2024-03-10 A Zero Trust Framework for Realization and Defense Against Generative AI Attacks in Power Grid. (22%)Md. Shirajum Munir; Sravanthi Proddatoori; Manjushree Muralidhara; Walid Saad; Zhu Han; Sachin Shetty 2024-03-09 Hard-label based Small Query Black-box Adversarial Attack. (99%)Jeonghwan Park; Paul Miller; Niall McLaughlin MirrorAttack: Backdoor Attack on 3D Point Cloud with a Distorting Mirror. (81%)Yuhao Bian; Shengjing Tian; Xiuping Liu IOI: Invisible One-Iteration Adversarial Attack on No-Reference Image- and Video-Quality Metrics. (78%)Ekaterina Shumitskaya; Anastasia Antsiferova; Dmitriy Vatolin Attacking Transformers with Feature Diversity Adversarial Perturbation. (70%)Chenxing Gao; Hang Zhou; Junqing Yu; YuTeng Ye; Jiale Cai; Junle Wang; Wei Yang 2024-03-08 Hide in Thicket: Generating Imperceptible and Rational Adversarial Perturbations on 3D Point Clouds. (99%)Tianrui Lou; Xiaojun Jia; Jindong Gu; Li Liu; Siyuan Liang; Bangyan He; Xiaochun Cao Exploring the Adversarial Frontier: Quantifying Robustness via Adversarial Hypervolume. (98%)Ping Guo; Cheng Gong; Xi Lin; Zhiyuan Yang; Qingfu Zhang Prepared for the Worst: A Learning-Based Adversarial Attack for Resilience Analysis of the ICP Algorithm. (93%)Ziyu Zhang; Johann Laconte; Daniil Lisus; Timothy D. Barfoot Adversarial Sparse Teacher: Defense Against Distillation-Based Model Stealing Attacks Using Adversarial Examples. (92%)Eda Yilmaz; Hacer Yalim Keles EVD4UAV: An Altitude-Sensitive Benchmark to Evade Vehicle Detection in UAV. (81%)Huiming Sun; Jiacheng Guo; Zibo Meng; Tianyun Zhang; Jianwu Fang; Yuewei Lin; Hongkai Yu Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. (70%)Machel Reid; Nikolay Savinov; Denis Teplyashin; Dmitry Lepikhin; Timothy Lillicrap; Jean-baptiste Alayrac; Radu Soricut; Angeliki Lazaridou; Orhan Firat; Julian Schrittwieser; Ioannis Antonoglou; Rohan Anil; Sebastian Borgeaud; Andrew Dai; Katie Millican; Ethan Dyer; Mia Glaese; Thibault Sottiaux; Benjamin Lee; Fabio Viola; Malcolm Reynolds; Yuanzhong Xu; James Molloy; Jilin Chen; Michael Isard; Paul Barham; Tom Hennigan; Ross McIlroy; Melvin Johnson; Johan Schalkwyk; Eli Collins; Eliza Rutherford; Erica Moreira; Kareem Ayoub; Megha Goel; Clemens Meyer; Gregory Thornton; Zhen Yang; Henryk Michalewski; Zaheer Abbas; Nathan Schucher; Ankesh Anand; Richard Ives; James Keeling; Karel Lenc; Salem Haykal; Siamak Shakeri; Pranav Shyam; Aakanksha Chowdhery; Roman Ring; Stephen Spencer; Eren Sezener; Luke Vilnis; Oscar Chang; Nobuyuki Morioka; George Tucker; Ce Zheng; Oliver Woodman; Nithya Attaluri; Tomas Kocisky; Evgenii Eltyshev; Xi Chen; Timothy Chung; Vittorio Selo; Siddhartha Brahma; Petko Georgiev; Ambrose Slone; Zhenkai Zhu; James Lottes; Siyuan Qiao; Ben Caine; Sebastian Riedel; Alex Tomala; Martin Chadwick; Juliette Love; Peter Choy; Sid Mittal; Neil Houlsby; Yunhao Tang; Matthew Lamm; Libin Bai; Qiao Zhang; Luheng He; Yong Cheng; Peter Humphreys; Yujia Li; Sergey Brin; Albin Cassirer; Yingjie Miao; Lukas Zilka; Taylor Tobin; Kelvin Xu; Lev Proleev; Daniel Sohn; Alberto Magni; Lisa Anne Hendricks; Isabel Gao; Santiago Ontañón; Oskar Bunyan; Nathan Byrd; Abhanshu Sharma; Biao Zhang; Mario Pinto; Rishika Sinha; Harsh Mehta; Dawei Jia; Sergi Caelles; Albert Webson; Alex Morris; Becca Roelofs; Yifan Ding; Robin Strudel; Xuehan Xiong; Marvin Ritter; Mostafa Dehghani; Rahma Chaabouni; Abhijit Karmarkar; Guangda Lai; Fabian Mentzer; Bibo Xu; YaGuang Li; Yujing Zhang; Tom Le Paine; Alex Goldin; Behnam Neyshabur; Kate Baumli; Anselm Levskaya; Michael Laskin; Wenhao Jia; Jack W. Rae; Kefan Xiao; Antoine He; Skye Giordano; Lakshman Yagati; Jean-Baptiste Lespiau; Paul Natsev; Sanjay Ganapathy; Fangyu Liu; Danilo Martins; Nanxin Chen; Yunhan Xu; Megan Barnes; Rhys May; Arpi Vezer; Junhyuk Oh; Ken Franko; Sophie Bridgers; Ruizhe Zhao; Boxi Wu; Basil Mustafa; Sean Sechrist; Emilio Parisotto; Thanumalayan Sankaranarayana Pillai; Chris Larkin; Chenjie Gu; Christina Sorokin; Maxim Krikun; Alexey Guseynov; Jessica Landon; Romina Datta; Alexander Pritzel; Phoebe Thacker; Fan Yang; Kevin Hui; Anja Hauth; Chih-Kuan Yeh; David Barker; Justin Mao-Jones; Sophia Austin; Hannah Sheahan; Parker Schuh; James Svensson; Rohan Jain; Vinay Ramasesh; Anton Briukhov; Da-Woon Chung; Glehn Tamara von; Christina Butterfield; Priya Jhakra; Matthew Wiethoff; Justin Frye; Jordan Grimstad; Beer Changpinyo; Charline Le Lan; Anna Bortsova; Yonghui Wu; Paul Voigtlaender; Tara Sainath; Charlotte Smith; Will Hawkins; Kris Cao; James Besley; Srivatsan Srinivasan; Mark Omernick; Colin Gaffney; Gabriela Surita; Ryan Burnell; Bogdan Damoc; Junwhan Ahn; Andrew Brock; Mantas Pajarskas; Anastasia Petrushkina; Seb Noury; Lorenzo Blanco; Kevin Swersky; Arun Ahuja; Thi Avrahami; Vedant Misra; Liedekerke Raoul de; Mariko Iinuma; Alex Polozov; Sarah York; George van den Driessche; Paul Michel; Justin Chiu; Rory Blevins; Zach Gleicher; Adrià Recasens; Alban Rrustemi; Elena Gribovskaya; Aurko Roy; Wiktor Gworek; Séb Arnold; Lisa Lee; James Lee-Thorp; Marcello Maggioni; Enrique Piqueras; Kartikeya Badola; Sharad Vikram; Lucas Gonzalez; Anirudh Baddepudi; Evan Senter; Jacob Devlin; James Qin; Michael Azzam; Maja Trebacz; Martin Polacek; Kashyap Krishnakumar; Shuo-yiin Chang; Matthew Tung; Ivo Penchev; Rishabh Joshi; Kate Olszewska; Carrie Muir; Mateo Wirth; Ale Jakse Hartman; Josh Newlan; Sheleem Kashem; Vijay Bolina; Elahe Dabir; Amersfoort Joost van; Zafarali Ahmed; James Cobon-Kerr; Aishwarya Kamath; Arnar Mar Hrafnkelsson; Le Hou; Ian Mackinnon; Alexandre Frechette; Eric Noland; Xiance Si; Emanuel Taropa; Dong Li; Phil Crone; Anmol Gulati; Sébastien Cevey; Jonas Adler; Ada Ma; David Silver; Simon Tokumine; Richard Powell; Stephan Lee; Michael Chang; Samer Hassan; Diana Mincu; Antoine Yang; Nir Levine; Jenny Brennan; Mingqiu Wang; Sarah Hodkinson; Jeffrey Zhao; Josh Lipschultz; Aedan Pope; Michael B. Chang; Cheng Li; Laurent El Shafey; Michela Paganini; Sholto Douglas; Bernd Bohnet; Fabio Pardo; Seth Odoom; Mihaela Rosca; Cicero Nogueira dos Santos; Kedar Soparkar; Arthur Guez; Tom Hudson; Steven Hansen; Chulayuth Asawaroengchai; Ravi Addanki; Tianhe Yu; Wojciech Stokowiec; Mina Khan; Justin Gilmer; Jaehoon Lee; Carrie Grimes Bostock; Keran Rong; Jonathan Caton; Pedram Pejman; Filip Pavetic; Geoff Brown; Vivek Sharma; Mario Lučić; Rajkumar Samuel; Josip Djolonga; Amol Mandhane; Lars Lowe Sjösund; Elena Buchatskaya; Elspeth White; Natalie Clay; Jiepu Jiang; Hyeontaek Lim; Ross Hemsley; Jane Labanowski; Cao Nicola De; David Steiner; Sayed Hadi Hashemi; Jacob Austin; Anita Gergely; Tim Blyth; Joe Stanton; Kaushik Shivakumar; Aditya Siddhant; Anders Andreassen; Carlos Araya; Nikhil Sethi; Rakesh Shivanna; Steven Hand; Ankur Bapna; Ali Khodaei; Antoine Miech; Garrett Tanzer; Andy Swing; Shantanu Thakoor; Zhufeng Pan; Zachary Nado; Stephanie Winkler; Dian Yu; Mohammad Saleh; Loren Maggiore; Iain Barr; Minh Giang; Thais Kagohara; Ivo Danihelka; Amit Marathe; Vladimir Feinberg; Mohamed Elhawaty; Nimesh Ghelani; Dan Horgan; Helen Miller; Lexi Walker; Richard Tanburn; Mukarram Tariq; Disha Shrivastava; Fei Xia; Chung-Cheng Chiu; Zoe Ashwood; Khuslen Baatarsukh; Sina Samangooei; Fred Alcober; Axel Stjerngren; Paul Komarek; Katerina Tsihlas; Anudhyan Boral; Ramona Comanescu; Jeremy Chen; Ruibo Liu; Dawn Bloxwich; Charlie Chen; Yanhua Sun; Fangxiaoyu Feng; Matthew Mauger; Xerxes Dotiwalla; Vincent Hellendoorn; Michael Sharman; Ivy Zheng; Krishna Haridasan; Gabe Barth-Maron; Craig Swanson; Dominika Rogozińska; Alek Andreev; Paul Kishan Rubenstein; Ruoxin Sang; Dan Hurt; Gamaleldin Elsayed; Renshen Wang; Dave Lacey; Anastasija Ilić; Yao Zhao; Lora Aroyo; Chimezie Iwuanyanwu; Vitaly Nikolaev; Balaji Lakshminarayanan; Sadegh Jazayeri; Raphaël Lopez Kaufman; Mani Varadarajan; Chetan Tekur; Doug Fritz; Misha Khalman; David Reitter; Kingshuk Dasgupta; Shourya Sarcar; Tina Ornduff; Javier Snaider; Fantine Huot; Johnson Jia; Rupert Kemp; Nejc Trdin; Anitha Vijayakumar; Lucy Kim; Christof Angermueller; Li Lao; Tianqi Liu; Haibin Zhang; David Engel; Somer Greene; Anaïs White; Jessica Austin; Lilly Taylor; Shereen Ashraf; Dangyi Liu; Maria Georgaki; Irene Cai; Yana Kulizhskaya; Sonam Goenka; Brennan Saeta; Kiran Vodrahalli; Christian Frank; Cesare Dario de; Brona Robenek; Harry Richardson; Mahmoud Alnahlawi; Christopher Yew; Priya Ponnapalli; Marco Tagliasacchi; Alex Korchemniy; Yelin Kim; Dinghua Li; Bill Rosgen; Zoe Ashwood; Kyle Levin; Jeremy Wiesner; Praseem Banzal; Praveen Srinivasan; Hongkun Yu; Çağlar Ünlü; David Reid; Zora Tung; Daniel Finchelstein; Ravin Kumar; Andre Elisseeff; Jin Huang; Ming Zhang; Rui Zhu; Ricardo Aguilar; Mai Giménez; Jiawei Xia; Olivier Dousse; Willi Gierke; Soheil Hassas Yeganeh; Damion Yates; Komal Jalan; Lu Li; Eri Latorre-Chimoto; Duc Dung Nguyen; Ken Durden; Praveen Kallakuri; Yaxin Liu; Matthew Johnson; Tomy Tsai; Alice Talbert; Jasmine Liu; Alexander Neitz; Chen Elkind; Marco Selvi; Mimi Jasarevic; Livio Baldini Soares; Albert Cui; Pidong Wang; Alek Wenjiao Wang; Xinyu Ye; Krystal Kallarackal; Lucia Loher; Hoi Lam; Josef Broder; Dan Holtmann-Rice; Nina Martin; Bramandia Ramadhana; Daniel Toyama; Mrinal Shukla; Sujoy Basu; Abhi Mohan; Nick Fernando; Noah Fiedel; Kim Paterson; Hui Li; Ankush Garg; Jane Park; DongHyun Choi; Diane Wu; Sankalp Singh; Zhishuai Zhang; Amir Globerson; Lily Yu; John Carpenter; Félix de Chaumont Quitry; Carey Radebaugh; Chu-Cheng Lin; Alex Tudor; Prakash Shroff; Drew Garmon; Dayou Du; Neera Vats; Han Lu; Shariq Iqbal; Alex Yakubovich; Nilesh Tripuraneni; James Manyika; Haroon Qureshi; Nan Hua; Christel Ngani; Maria Abi Raad; Hannah Forbes; Anna Bulanova; Jeff Stanway; Mukund Sundararajan; Victor Ungureanu; Colton Bishop; Yunjie Li; Balaji Venkatraman; Bo Li; Chloe Thornton; Salvatore Scellato; Nishesh Gupta; Yicheng Wang; Ian Tenney; Xihui Wu; Ashish Shenoy; Gabriel Carvajal; Diana Gage Wright; Ben Bariach; Zhuyun Xiao; Peter Hawkins; Sid Dalmia; Clement Farabet; Pedro Valenzuela; Quan Yuan; Chris Welty; Ananth Agarwal; Mia Chen; Wooyeol Kim; Brice Hulse; Nandita Dukkipati; Adam Paszke; Andrew Bolt; Elnaz Davoodi; Kiam Choo; Jennifer Beattie; Jennifer Prendki; Harsha Vashisht; Rebeca Santamaria-Fernandez; Luis C. Cobo; Jarek Wilkiewicz; David Madras; Ali Elqursh; Grant Uy; Kevin Ramirez; Matt Harvey; Tyler Liechty; Heiga Zen; Jeff Seibert; Clara Huiyi Hu; Mohamed Elhawaty; Andrey Khorlin; Maigo Le; Asaf Aharoni; Megan Li; Lily Wang; Sandeep Kumar; Alejandro Lince; Norman Casagrande; Jay Hoover; Dalia El Badawy; David Soergel; Denis Vnukov; Matt Miecnikowski; Jiri Simsa; Anna Koop; Praveen Kumar; Thibault Sellam; Daniel Vlasic; Samira Daruki; Nir Shabat; John Zhang; Guolong Su; Jiageng Zhang; Jeremiah Liu; Yi Sun; Evan Palmer; Alireza Ghaffarkhah; Xi Xiong; Victor Cotruta; Michael Fink; Lucas Dixon; Ashwin Sreevatsa; Adrian Goedeckemeyer; Alek Dimitriev; Mohsen Jafari; Remi Crocker; Nicholas FitzGerald; Aviral Kumar; Sanjay Ghemawat; Ivan Philips; Frederick Liu; Yannie Liang; Rachel Sterneck; Alena Repina; Marcus Wu; Laura Knight; Marin Georgiev; Hyo Lee; Harry Askham; Abhishek Chakladar; Annie Louis; Carl Crous; Hardie Cate; Dessie Petrova; Michael Quinn; Denese Owusu-Afriyie; Achintya Singhal; Nan Wei; Solomon Kim; Damien Vincent; Milad Nasr; Christopher A. Choquette-Choo; Reiko Tojo; Shawn Lu; Diego de Las Casas; Yuchung Cheng; Tolga Bolukbasi; Katherine Lee; Saaber Fatehi; Rajagopal Ananthanarayanan; Miteyan Patel; Charbel Kaed; Jing Li; Jakub Sygnowski; Shreyas Rammohan Belle; Zhe Chen; Jaclyn Konzelmann; Siim Põder; Roopal Garg; Vinod Koverkathu; Adam Brown; Chris Dyer; Rosanne Liu; Azade Nova; Jun Xu; Slav Petrov; Demis Hassabis; Koray Kavukcuoglu; Jeffrey Dean; Oriol Vinyals The Impact of Quantization on the Robustness of Transformer-based Text Classifiers. (45%)Seyed Parsa Neshaei; Yasaman Boreshban; Gholamreza Ghassem-Sani; Seyed Abolghasem Mirroshandel 2024-03-07 Defending Against Unforeseen Failure Modes with Latent Adversarial Training. (83%)Stephen Casper; Lennart Schulze; Oam Patel; Dylan Hadfield-Menell Fooling Neural Networks for Motion Forecasting via Adversarial Attacks. (33%)Edgar Medina; Leyong Loh Automatic and Universal Prompt Injection Attacks against Large Language Models. (31%)Xiaogeng Liu; Zhiyuan Yu; Yizhe Zhang; Ning Zhang; Chaowei Xiao ObjectCompose: Evaluating Resilience of Vision-Based Models on Object-to-Background Compositional Changes. (31%)Hashmat Shadab Malik; Muhammad Huzaifa; Muzammal Naseer; Salman Khan; Fahad Shahbaz Khan Cell reprogramming design by transfer learning of functional transcriptional networks. (1%)Thomas P. Wytock; Adilson E. Motter Towards Robustness Analysis of E-Commerce Ranking System. (1%)Ningfei Wang; Yupin Huang; Han Cheng; Jiri Gesi; Xiaojie Wang; Vivek Mittal 2024-03-06 Adversarial Infrared Geometry: Using Geometry to Perform Adversarial Attack against Infrared Pedestrian Detectors. (99%)Kalibinuer Tiliwalidi Improving Adversarial Training using Vulnerability-Aware Perturbation Budget. (99%)Olukorede Fakorede; Modeste Atsague; Jin Tian Effect of Ambient-Intrinsic Dimension Gap on Adversarial Vulnerability. (92%)Rajdeep Haldar; Yue Xing; Qifan Song Belief-Enriched Pessimistic Q-Learning against Adversarial State Perturbations. (16%)Xiaolin Sun; Zizhan Zheng On the Effectiveness of Distillation in Mitigating Backdoors in Pre-trained Encoder. (2%)Tingxu Han; Shenghan Huang; Ziqi Ding; Weisong Sun; Yebo Feng; Chunrong Fang; Jun Li; Hanwei Qian; Cong Wu; Quanjun Zhang; Yang Liu; Zhenyu Chen Verified Training for Counterfactual Explanation Robustness under Data Shift. (2%)Anna P. Meyer; Yuhao Zhang; Aws Albarghouthi; Loris D'Antoni 2024-03-05 Towards Robust Federated Learning via Logits Calibration on Non-IID Data. (99%)Yu Qiao; Apurba Adhikary; Chaoning Zhang; Choong Seon Hong Mitigating Label Flipping Attacks in Malicious URL Detectors Using Ensemble Trees. (96%)Ehsan Nowroozi; Nada Jadalla; Samaneh Ghelichkhani; Alireza Jolfaei Minimum Topology Attacks for Graph Neural Networks. (83%)Mengmei Zhang; Xiao Wang; Chuan Shi; Lingjuan Lyu; Tianchi Yang; Junping Du Federated Learning Under Attack: Exposing Vulnerabilities through Data Poisoning Attacks in Computer Networks. (82%)Ehsan Nowroozi; Imran Haider; Rahim Taheri; Mauro Conti A general approach to enhance the survivability of backdoor attacks by decision path coupling. (68%)Yufei Zhao; Dingji Wang; Bihuan Chen; Ziqian Chen; Xin Peng Robust Federated Learning Mitigates Client-side Training Data Distribution Inference Attacks. (61%)Yichang Xu; Ming Yin; Minghong Fang; Neil Zhenqiang Gong Uplift Modeling for Target User Attacks on Recommender Systems. (12%)Wenjie Wang; Changsheng Wang; Fuli Feng; Wentao Shi; Daizong Ding; Tat-Seng Chua FLGuard: Byzantine-Robust Federated Learning via Ensemble of Contrastive Models. (11%)Younghan Lee; Yungi Cho; Woorim Han; Ho Bae; Yunheung Paek InjecAgent: Benchmarking Indirect Prompt Injections in Tool-Integrated Large Language Model Agents. (11%)Qiusi Zhan; Zhixiang Liang; Zifan Ying; Daniel Kang XAI-Based Detection of Adversarial Attacks on Deepfake Detectors. (8%)Ben Pinhasov; Raz Lapid; Rony Ohayon; Moshe Sipper; Yehudit Aperstein 2024-03-04 Robustness Bounds on the Successful Adversarial Examples: Theory and Practice. (99%)Hiroaki Maeshima; Akira Otsuka One Prompt Word is Enough to Boost Adversarial Robustness for Pre-trained Vision-Language Models. (99%)Lin Li; Haoyan Guan; Jianing Qiu; Michael Spratling Improving the Robustness of Object Detection and Classification AI models against Adversarial Patch Attacks. (99%)Roie Kazoom; Raz Birman; Ofer Hadar COMMIT: Certifying Robustness of Multi-Sensor Fusion Systems against Semantic Attacks. (96%)Zijian Huang; Wenda Chu; Linyi Li; Chejian Xu; Bo Li Inf2Guard: An Information-Theoretic Framework for Learning Privacy-Preserving Representations against Inference Attacks. (26%)Sayedeh Leila Noorbakhsh; Binghui Zhang; Yuan Hong; Binghui Wang BSDP: Brain-inspired Streaming Dual-level Perturbations for Online Open World Object Detection. (16%)Yu Chen; Liyan Ma; Liping Jing; Jian Yu Mirage: Defense against CrossPath Attacks in Software Defined Networks. (3%)Shariq Murtuza; Krishna Asawa 2024-03-03 GuardT2I: Defending Text-to-Image Models from Adversarial Prompts. (9%)Yijun Yang; Ruiyuan Gao; Xiao Yang; Jianyuan Zhong; Qiang Xu 2024-03-02 SAR-AE-SFP: SAR Imagery Adversarial Example in Real Physics domain with Target Scattering Feature Parameters. (99%)Jiahao Cui; Jiale Duan; Binyan Luo; Hang Cao; Wang Guo; Haifeng Li Inexact Unlearning Needs More Careful Evaluations to Avoid a False Sense of Privacy. (68%)Jamie Hayes; Ilia Shumailov; Eleni Triantafillou; Amr Khalifa; Nicolas Papernot Breaking Down the Defenses: A Comparative Survey of Attacks on Large Language Models. (56%)Arijit Ghosh Chowdhury; Md Mofijul Islam; Vaibhav Kumar; Faysal Hossain Shezan; Vaibhav Kumar; Vinija Jain; Aman Chadha Adversarial Testing for Visual Grounding via Image-Aware Property Reduction. (11%)Zhiyuan Chang; Mingyang Li; Junjie Wang; Cheng Li; Boyu Wu; Fanjiang Xu; Qing Wang Query Recovery from Easy to Hard: Jigsaw Attack against SSE. (2%)Hao Nie; Wei Wang; Peng Xu; Xianglong Zhang; Laurence T. Yang; Kaitai Liang 2024-03-01 Robust Deep Reinforcement Learning Through Adversarial Attacks and Training : A Survey. (91%)Lucas Schott; Josephine Delas; Hatem Hajri; Elies Gherbi; Reda Yaich; Nora Boulahia-Cuppens; Frederic Cuppens; Sylvain Lamprier Resilience of Entropy Model in Distributed Neural Networks. (67%)Milin Zhang; Mohammad Abdi; Shahriar Rifat; Francesco Restuccia Attacking Delay-based PUFs with Minimal Adversary Model. (45%)Hongming Fei; Owen Millwood; Prosanta Gope; Jack Miskelly; Biplab Sikdar 2024-02-29 Unraveling Adversarial Examples against Speaker Identification -- Techniques for Attack Detection and Victim Model Classification. (99%)Sonal Joshi; Thomas Thebaud; Jesús Villalba; Najim Dehak How to Train your Antivirus: RL-based Hardening through the Problem-Space. (99%)Jacopo Cortellazzi; Ilias Tsingenopoulos; Branislav Bošanský; Simone Aonzo; Davy Preuveneers; Wouter Joosen; Fabio Pierazzi; Lorenzo Cavallaro On Robustness and Generalization of ML-Based Congestion Predictors to Valid and Imperceptible Perturbations. (88%)Chester Holtz; Yucheng Wang; Chung-Kuan Cheng; Bill Lin Pointing out the Shortcomings of Relation Extraction Models with Semantically Motivated Adversarials. (76%)Gennaro Nolano; Moritz Blum; Basil Ell; Philipp Cimiano Assessing Visually-Continuous Corruption Robustness of Neural Networks Relative to Human Performance. (38%)Huakun Shen; Boyue Caroline Hu; Krzysztof Czarnecki; Lina Marsso; Marsha Chechik Verification of Neural Networks' Global Robustness. (38%)Anan Kabaha; Dana Drachsler-Cohen Here's a Free Lunch: Sanitizing Backdoored Models with Model Merge. (2%)Ansh Arora; Xuanli He; Maximilian Mozes; Srinibas Swain; Mark Dras; Qiongkai Xu Gradient Cuff: Detecting Jailbreak Attacks on Large Language Models by Exploring Refusal Loss Landscapes. (1%)Xiaomeng Hu; Pin-Yu Chen; Tsung-Yi Ho 2024-02-28 Enhancing the "Immunity" of Mixture-of-Experts Networks for Adversarial Defense. (99%)Qiao Han; yong huang; xinling Guo; Yiteng Zhai; Yu Qin; Yao Yang MPAT: Building Robust Deep Neural Networks against Textual Adversarial Attacks. (99%)Fangyuan Zhang; Huichi Zhou; Shuangjiao Li; Hongtao Wang Catastrophic Overfitting: A Potential Blessing in Disguise. (98%)Mengnan Zhao; Lihe Zhang; Yuqiu Kong; Baocai Yin Living-off-The-Land Reverse-Shell Detection by Informed Data Augmentation. (76%)Dmitrijs Trizna; Luca Demetrio; Battista Biggio; Fabio Roli A New Era in LLM Security: Exploring Security Concerns in Real-World LLM-based Systems. (64%)Fangzhou Wu; Ning Zhang; Somesh Jha; Patrick McDaniel; Chaowei Xiao Making Them Ask and Answer: Jailbreaking Large Language Models in Few Queries via Disguise and Reconstruction. (33%)Tong Liu; Yingjie Zhang; Zhe Zhao; Yinpeng Dong; Guozhu Meng; Kai Chen Out-of-Distribution Detection using Neural Activation Prior. (1%)Weilin Wan; Weizhong Zhang; Cheng Jin 2024-02-27 Robustness-Congruent Adversarial Training for Secure Machine Learning Model Updates. (99%)Daniele Angioni; Luca Demetrio; Maura Pintor; Luca Oneto; Davide Anguita; Battista Biggio; Fabio Roli Extreme Miscalibration and the Illusion of Adversarial Robustness. (99%)Vyas Raina; Samson Tan; Volkan Cevher; Aditya Rawal; Sheng Zha; George Karypis Black-box Adversarial Attacks Against Image Quality Assessment Models. (99%)Yu Ran; Ao-Xiang Zhang; Mingjie Li; Weixuan Tang; Yuan-Gen Wang Enhancing Tracking Robustness with Auxiliary Adversarial Defense Networks. (99%)Zhewei Wu; Ruilong Yu; Qihe Liu; Shuying Cheng; Shilin Qiu; Shijie Zhou LLM-Resistant Math Word Problem Generation via Adversarial Attacks. (87%)Roy Xie; Chengxuan Huang; Junlin Wang; Bhuwan Dhingra Breaking the Black-Box: Confidence-Guided Model Inversion Attack for Distribution Shift. (83%)Xinhao Liu; Yingzhao Jiang; Zetao Lin Model X-ray:Detect Backdoored Models via Decision Boundary. (67%)Yanghao Su; Jie Zhang; Ting Xu; Tianwei Zhang; Weiming Zhang; Nenghai Yu Towards Fairness-Aware Adversarial Learning. (11%)Yanghao Zhang; Tianle Zhang; Ronghui Mu; Xiaowei Huang; Wenjie Ruan Time-Restricted Double-Spending Attack on PoW-based Blockchains. (1%)Yiming Jiang; Jiangfan Zhang 2024-02-26 Improving the JPEG-resistance of Adversarial Attacks on Face Recognition by Interpolation Smoothing. (99%)Kefu Guo; Fengfan Zhou; Hefei Ling; Ping Li; Hui Liu Improving behavior based authentication against adversarial attack using XAI. (99%)Dong Qin; George Amariucai; Daji Qiao; Yong Guan Adversarial example soups: averaging multiple adversarial examples improves transferability without increasing additional generation time. (99%)Bo Yang; Hengwei Zhang; Chenwei Li; Jindong Wang A Curious Case of Remarkable Resilience to Gradient Attacks via Fully Convolutional and Differentiable Front End with a Skip Connection. (98%)Leonid Boytsov; Ameya Joshi; Filipe Condessa Adversarial Perturbations of Physical Signals. (92%)Robert L. Bassett; Dellen Austin Van; Anthony P. Austin Unveiling Vulnerability of Self-Attention. (87%)Khai Jiet Liong; Hongqiu Wu; Hai Zhao Edge Detectors Can Make Deep Convolutional Neural Networks More Robust. (83%)Jin Ding; Jie-Chao Zhao; Yong-Zhi Sun; Ping Tan; Jia-Wei Wang; Ji-En Ma; You-Tong Fang Investigating Deep Watermark Security: An Adversarial Transferability Perspective. (64%)Biqing Qi; Junqi Gao; Yiang Luo; Jianxing Liu; Ligang Wu; Bowen Zhou Defending LLMs against Jailbreaking Attacks via Backtranslation. (33%)Yihan Wang; Zhouxing Shi; Andrew Bai; Cho-Jui Hsieh Pandora's White-Box: Increased Training Data Leakage in Open LLMs. (13%)Jeffrey G. Wang; Jason Wang; Marvin Li; Seth Neel WIPI: A New Web Threat for LLM-Driven Web Agents. (8%)Fangzhou Wu; Shutong Wu; Yulong Cao; Chaowei Xiao RoCoIns: Enhancing Robustness of Large Language Models through Code-Style Instructions. (4%)Yuansen Zhang; Xiao Wang; Zhiheng Xi; Han Xia; Tao Gui; Qi Zhang; Xuanjing Huang An Innovative Information Theory-based Approach to Tackle and Enhance The Transparency in Phishing Detection. (1%)Van Nguyen; Tingmin Wu; Xingliang Yuan; Marthie Grobler; Surya Nepal; Carsten Rudolph 2024-02-25 From Noise to Clarity: Unraveling the Adversarial Suffix of Large Language Model Attacks via Translation of Text Embeddings. (98%)Hao Wang; Hao Li; Minlie Huang; Lei Sha An Adversarial Robustness Benchmark for Enterprise Network Intrusion Detection. (92%)João Vitorino; Miguel Silva; Eva Maia; Isabel Praça Defending Large Language Models against Jailbreak Attacks via Semantic Smoothing. (76%)Jiabao Ji; Bairu Hou; Alexander Robey; George J. Pappas; Hamed Hassani; Yang Zhang; Eric Wong; Shiyu Chang Adversarial-Robust Transfer Learning for Medical Imaging via Domain Assimilation. (73%)Xiaohui Chen; Tie Luo Evaluating Robustness of Generative Search Engine on Adversarial Factual Questions. (13%)Xuming Hu; Xiaochuan Li; Junzhe Chen; Yinghui Li; Yangning Li; Xiaoguang Li; Yasheng Wang; Qun Liu; Lijie Wen; Philip S. Yu; Zhijiang Guo DrAttack: Prompt Decomposition and Reconstruction Makes Powerful LLM Jailbreakers. (2%)Xirui Li; Ruochen Wang; Minhao Cheng; Tianyi Zhou; Cho-Jui Hsieh m2mKD: Module-to-Module Knowledge Distillation for Modular Transformers. (1%)Ka Man Lo; Yiming Liang; Wenyu Du; Yuantao Fan; Zili Wang; Wenhao Huang; Lei Ma; Jie Fu 2024-02-24 PRP: Propagating Universal Perturbations to Attack Large Language Model Guard-Rails. (87%)Neal Mangaokar; Ashish Hooda; Jihye Choi; Shreyas Chandrashekaran; Kassem Fawaz; Somesh Jha; Atul Prakash LLMs Can Defend Themselves Against Jailbreaking in a Practical Manner: A Vision Paper. (86%)Daoyuan Wu; Shuai Wang; Yang Liu; Ning Liu RAUCA: A Novel Physical Adversarial Attack on Vehicle Detectors via Robust and Accurate Camouflage Generation. (82%)Jiawei Zhou; Linye Lyu; Daojing He; Yu Li Towards Robust Image Stitching: An Adaptive Resistance Learning against Compatible Attacks. (76%)Zhiying Jiang; Xingyuan Li; Jinyuan Liu; Xin Fan; Risheng Liu Optimal Zero-Shot Detector for Multi-Armed Attacks. (50%)Federica Granese; Marco Romanelli; Pablo Piantanida Sparse MeZO: Less Parameters for Better Performance in Zeroth-Order LLM Fine-Tuning. (1%)Yong Liu; Zirui Zhu; Chaoyu Gong; Minhao Cheng; Cho-Jui Hsieh; Yang You 2024-02-23 Distilling Adversarial Robustness Using Heterogeneous Teachers. (99%)Jieren Deng; Aaron Palmer; Rigel Mahmood; Ethan Rathbun; Jinbo Bi; Kaleel Mahmood; Derek Aguiar Fast Adversarial Attacks on Language Models In One GPU Minute. (98%)Vinu Sankar Sadasivan; Shoumik Saha; Gaurang Sriramanan; Priyatham Kattakinda; Atoosa Chegini; Soheil Feizi A Robust Defense against Adversarial Attacks on Deep Learning-based Malware Detectors via (De)Randomized Smoothing. (98%)Daniel Gibert; Giulio Zizzo; Quan Le; Jordi Planes ProTIP: Probabilistic Robustness Verification on Text-to-Image Diffusion Models against Stochastic Perturbation. (93%)Yi Zhang; Yun Tang; Wenjie Ruan; Xiaowei Huang; Siddartha Khastgir; Paul Jennings; Xingyu Zhao On the Duality Between Sharpness-Aware Minimization and Adversarial Training. (92%)Yihao Zhang; Hangzhou He; Jingyu Zhu; Huanran Chen; Yifei Wang; Zeming Wei Low-Frequency Black-Box Backdoor Attack via Evolutionary Algorithm. (87%)Yanqi Qiao; Dazhuang Liu; Rui Wang; Kaitai Liang Deep Networks Always Grok and Here is Why. (76%)Ahmed Imtiaz Humayun; Randall Balestriero; Richard Baraniuk BSPA: Exploring Black-box Stealthy Prompt Attacks against Image Generators. (67%)Yu Tian; Xiao Yang; Yinpeng Dong; Heming Yang; Hang Su; Jun Zhu Reinforcement Learning-Based Approaches for Enhancing Security and Resilience in Smart Control: A Survey on Attack and Defense Methods. (61%)Zheyu Zhang Break the Breakout: Reinventing LM Defense Against Jailbreak Attacks with Self-Refinement. (5%)Heegyu Kim; Sehyun Yuk; Hyunsouk Cho Prime+Retouch: When Cache is Locked and Leaked. (2%)Jaehyuk Lee; Fan Sang; Taesoo Kim 2024-02-22 SoK: Analyzing Adversarial Examples: A Framework to Study Adversary Knowledge. (99%)Lucas Fenaux; Florian Kerschbaum Rethinking Invariance Regularization in Adversarial Training to Improve Robustness-Accuracy Trade-off. (98%)Futa Waseda; Isao Echizen Noise-BERT: A Unified Perturbation-Robust Framework with Noise Alignment Pre-training for Noisy Slot Filling Task. (83%)Jinxu Zhao; Guanting Dong; Yueyan Qiu; Tingfeng Hui; Xiaoshuai Song; Daichi Guo; Weiran Xu Stop Reasoning! When Multimodal LLMs with Chain-of-Thought Reasoning Meets Adversarial Images. (81%)Zefeng Wang; Zhen Han; Shuo Chen; Fan Xue; Zifeng Ding; Xun Xiao; Volker Tresp; Philip Torr; Jindong Gu Mitigating Fine-tuning Jailbreak Attack with Backdoor Enhanced Alignment. (75%)Jiongxiao Wang; Jiazhao Li; Yiquan Li; Xiangyu Qi; Junjie Hu; Yixuan Li; Patrick McDaniel; Muhao Chen; Bo Li; Chaowei Xiao Getting Serious about Humor: Crafting Humor Datasets with Unfunny Large Language Models. (26%)Zachary Horvitz; Jingru Chen; Rahul Aditya; Harshvardhan Srivastava; Robert West; Zhou Yu; Kathleen McKeown 2024-02-21 AttackGNN: Red-Teaming GNNs in Hardware Security Using Reinforcement Learning. (99%)Vasudev Gohil; Satwik Patnaik; Dileep Kalathil; Jeyavijayan Rajendran A Simple and Yet Fairly Effective Defense for Graph Neural Networks. (98%)Sofiane Ennadir; Yassine Abbahaddou; Johannes F. Lutzeyer; Michalis Vazirgiannis; Henrik Boström Adversarial Purification and Fine-tuning for Robust UDC Image Restoration. (98%)Zhenbo Song; Zhenyuan Zhang; Kaihao Zhang; Wenhan Luo; Zhaoxin Fan; Jianfeng Lu Robustness of Deep Neural Networks for Micro-Doppler Radar Classification. (80%)Mikolaj Czerkawski; Carmine Clemente; Craig MichieCraig Michie; Christos Tachtatzis Is LLM-as-a-Judge Robust? Investigating Universal Adversarial Attacks on Zero-shot LLM Assessment. (75%)Vyas Raina; Adian Liusie; Mark Gales Flexible Physical Camouflage Generation Based on a Differential Approach. (38%)Yang Li; Wenyi Tan; Chenxing Zhao; Shuangju Zhou; Xinkai Liang; Quan Pan VL-Trojan: Multimodal Instruction Backdoor Attacks against Autoregressive Visual Language Models. (10%)Jiawei Liang; Siyuan Liang; Man Luo; Aishan Liu; Dongchen Han; Ee-Chien Chang; Xiaochun Cao Semantic Mirror Jailbreak: Genetic Algorithm Based Jailbreak Prompts Against Open-source LLMs. (8%)Xiaoxia Li; Siyuan Liang; Jiyi Zhang; Han Fang; Aishan Liu; Ee-Chien Chang Coercing LLMs to do and reveal (almost) anything. (4%)Jonas Geiping; Alex Stein; Manli Shu; Khalid Saifullah; Yuxin Wen; Tom Goldstein T-Stitch: Accelerating Sampling in Pre-Trained Diffusion Models with Trajectory Stitching. (1%)Zizheng Pan; Bohan Zhuang; De-An Huang; Weili Nie; Zhiding Yu; Chaowei Xiao; Jianfei Cai; Anima Anandkumar 2024-02-20 QuanTest: Entanglement-Guided Testing of Quantum Neural Network Systems. (92%)Jinjing Shi; Zimeng Xiao; Heyuan Shi; Yu Jiang; Xuelong Li Defending Jailbreak Prompts via In-Context Adversarial Game. (76%)Yujun Zhou; Yufei Han; Haomin Zhuang; Taicheng Guo; Kehan Guo; Zhenwen Liang; Hongyan Bao; Xiangliang Zhang Round Trip Translation Defence against Large Language Model Jailbreaking Attacks. (74%)Canaan Yung; Hadi Mohaghegh Dolatabadi; Sarah Erfani; Christopher Leckie Investigating the Impact of Model Instability on Explanations and Uncertainty. (69%)Sara Vera Marjanović; Isabelle Augenstein; Christina Lioma LLM Jailbreak Attack versus Defense Techniques -- A Comprehensive Study. (54%)Zihao Xu; Yi Liu; Gelei Deng; Yuekang Li; Stjepan Picek Learning to Poison Large Language Models During Instruction Tuning. (13%)Yao Qiang; Xiangyu Zhou; Saleh Zare Zade; Mohammad Amin Roshani; Douglas Zytko; Dongxiao Zhu Stealthy Adversarial Attacks on Stochastic Multi-Armed Bandits. (3%)Zhiwei Wang; Huazheng Wang; Hongning Wang RITFIS: Robust input testing framework for LLMs-based intelligent software. (1%)Mingxuan Xiao; Yan Xiao; Hai Dong; Shunhui Ji; Pengcheng Zhang 2024-02-19 Query-Based Adversarial Prompt Generation. (99%)Jonathan Hayase; Ema Borevkovic; Nicholas Carlini; Florian Tramèr; Milad Nasr Adversarial Feature Alignment: Balancing Robustness and Accuracy in Deep Learning via Adversarial Training. (99%)Leo Hyun Park; Jaeuk Kim; Myung Gyo Oh; Jaewoo Park; Taekyoung Kwon An Adversarial Approach to Evaluating the Robustness of Event Identification Models. (98%)Obai Bahwal; Oliver Kosut; Lalitha Sankar AICAttack: Adversarial Image Captioning Attack with Attention-Based Optimization. (98%)Jiyao Li; Mingze Ni; Yifei Dong; Tianqing Zhu; Wei Liu Beyond Worst-case Attacks: Robust RL with Adaptive Defense via Non-dominated Policies. (97%)Xiangyu Liu; Chenghao Deng; Yanchao Sun; Yongyuan Liang; Furong Huang Stealing the Invisible: Unveiling Pre-Trained CNN Models through Adversarial Examples and Timing Side-Channels. (92%)Shubhi Shukla; Manaar Alam; Pabitra Mitra; Debdeep Mukhopadhyay Attacks on Node Attributes in Graph Neural Networks. (83%)Ying Xu; Michael Lanier; Anindya Sarkar; Yevgeniy Vorobeychik Indiscriminate Data Poisoning Attacks on Pre-trained Feature Extractors. (68%)Yiwei Lu; Matthew Y. R. Yang; Gautam Kamath; Yaoliang Yu Self-Guided Robust Graph Structure Refinement. (67%)Yeonjun In; Kanghoon Yoon; Kibum Kim; Kijung Shin; Chanyoung Park Robust CLIP: Unsupervised Adversarial Fine-Tuning of Vision Embeddings for Robust Large Vision-Language Models. (50%)Christian Schlarmann; Naman Deep Singh; Francesco Croce; Matthias Hein Defending Against Weight-Poisoning Backdoor Attacks for Parameter-Efficient Fine-Tuning. (15%)Shuai Zhao; Leilei Gan; Luu Anh Tuan; Jie Fu; Lingjuan Lyu; Meihuizi Jia; Jinming Wen Amplifying Training Data Exposure through Fine-Tuning with Pseudo-Labeled Memberships. (1%)Myung Gyo Oh; Hong Eun Ahn; Leo Hyun Park; Taekyoung Kwon 2024-02-18 Evaluating Adversarial Robustness of Low dose CT Recovery. (92%)Kanchana Vaishnavi Gandikota; Paramanand Chandramouli; Hannah Droege; Michael Moeller A Curious Case of Searching for the Correlation between Training Data and Adversarial Robustness of Transformer Textual Models. (92%)Cuong Dang; Dung D. Le; Thai Le Evaluating Efficacy of Model Stealing Attacks and Defenses on Quantum Neural Networks. (83%)Satwik Kundu; Debarshi Kundu; Swaroop Ghosh The Effectiveness of Random Forgetting for Robust Generalization. (75%)Vijaya Raghavan T Ramkumar; Bahram Zonooz; Elahe Arani Poisoned Forgery Face: Towards Backdoor Attacks on Face Forgery Detection. (26%)Jiawei Liang; Siyuan Liang; Aishan Liu; Xiaojun Jia; Junhao Kuang; Xiaochun Cao Poisoning Federated Recommender Systems with Fake Users. (5%)Ming Yin; Yichang Xu; Minghong Fang; Neil Zhenqiang Gong SPML: A DSL for Defending Language Models Against Prompt Attacks. (1%)Reshabh K Sharma; Vinayak Gupta; Dan Grossman Teacher as a Lenient Expert: Teacher-Agnostic Data-Free Knowledge Distillation. (1%)Hyunjune Shin; Dong-Wan Choi 2024-02-17 Maintaining Adversarial Robustness in Continuous Learning. (75%)Xiaolei Ru; Xiaowei Cao; Zijia Liu; Jack Murdoch Moore; Xin-Ya Zhang; Xia Zhu; Wenjia Wei; Gang Yan Be Persistent: Towards a Unified Solution for Mitigating Shortcuts in Deep Learning. (22%)Hadi M. Dolatabadi; Sarah M. Erfani; Christopher Leckie Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents. (2%)Wenkai Yang; Xiaohan Bi; Yankai Lin; Sishuo Chen; Jie Zhou; Xu Sun VoltSchemer: Use Voltage Noise to Manipulate Your Wireless Charger. (2%)Zihao Zhan; Yirui Yang; Haoqi Shan; Hanqiu Wang; Yier Jin; Shuo Wang 2024-02-16 DART: A Principled Approach to Adversarially Robust Unsupervised Domain Adaptation. (99%)Yunjuan Wang; Hussein Hazimeh; Natalia Ponomareva; Alexey Kurakin; Ibrahim Hammoud; Raman Arora Theoretical Understanding of Learning from Adversarial Perturbations. (98%)Soichiro Kumano; Hiroshi Kera; Toshihiko Yamasaki Zero-shot sampling of adversarial entities in biomedical question answering. (92%)R. Patrick Xian; Alex J. Lee; Vincent Wang; Qiming Cui; Russell Ro; Reza Abbasi-Asl VQAttack: Transferable Adversarial Attacks on Visual Question Answering via Pre-trained Models. (92%)Ziyi Yin; Muchao Ye; Tianrong Zhang; Jiaqi Wang; Han Liu; Jinghui Chen; Ting Wang; Fenglong Ma The AI Security Pyramid of Pain. (47%)Chris M. Ward; Josh Harguess; Julia Tao; Daniel Christman; Paul Spicer; Mike Tan AIM: Automated Input Set Minimization for Metamorphic Security Testing. (2%)Nazanin Bayati Chaleshtari; Yoann Marquer; Fabrizio Pastore; Lionel C. Briand Universal Prompt Optimizer for Safe Text-to-Image Generation. (1%)Zongyu Wu; Hongcheng Gao; Yueze Wang; Xiang Zhang; Suhang Wang 2024-02-15 Camouflage is all you need: Evaluating and Enhancing Language Model Robustness Against Camouflage Adversarial Attacks. (62%)Álvaro Huertas-García; Alejandro Martín; Javier Huertas-Tato; David Camacho On the Safety Concerns of Deploying LLMs/VLMs in Robotics: Highlighting the Risks and Vulnerabilities. (31%)Xiyang Wu; Ruiqi Xian; Tianrui Guan; Jing Liang; Souradip Chakraborty; Fuxiao Liu; Brian Sadler; Dinesh Manocha; Amrit Singh Bedi Backdoor Attack against One-Class Sequential Anomaly Detection Models. (9%)He Cheng; Shuhan Yuan A Trembling House of Cards? Mapping Adversarial Attacks against Language Agents. (5%)Lingbo Mo; Zeyi Liao; Boyuan Zheng; Yu Su; Chaowei Xiao; Huan Sun FedRDF: A Robust and Dynamic Aggregation Function against Poisoning Attacks in Federated Learning. (3%)Enrique Mármol Campos; Aurora González Vidal; José Luis Hernández Ramos; Antonio Skarmeta Quantum-Inspired Analysis of Neural Network Vulnerabilities: The Role of Conjugate Variables in System Attacks. (1%)Jun-Jie Zhang; Deyu Meng 2024-02-14 Exploring the Adversarial Capabilities of Large Language Models. (98%)Lukas Struppek; Minh Hieu Le; Dominik Hintersdorf; Kristian Kersting PAL: Proxy-Guided Black-Box Attack on Large Language Models. (92%)Chawin Sitawarin; Norman Mu; David Wagner; Alexandre Araujo Only My Model On My Data: A Privacy Preserving Approach Protecting one Model and Deceiving Unauthorized Black-Box Models. (92%)Weiheng Chai; Brian Testa; Huantao Ren; Asif Salekin; Senem Velipasalar How Secure Are Large Language Models (LLMs) for Navigation in Urban Environments? (80%)Congcong Wen; Jiazhao Liang; Shuaihang Yuan; Hao Huang; Yi Fang Review-Incorporated Model-Agnostic Profile Injection Attacks on Recommender Systems. (76%)Shiyi Yang; Lina Yao; Chen Wang; Xiwei Xu; Liming Zhu Attacking Large Language Models with Projected Gradient Descent. (67%)Simon Geisler; Tom Wollschläger; M. H. I. Abdalla; Johannes Gasteiger; Stephan Günnemann Detecting Adversarial Spectrum Attacks via Distance to Decision Boundary Statistics. (47%)Wenwei Zhao; Xiaowen Li; Shangqing Zhao; Jie Xu; Yao Liu; Zhuo Lu SafeDecoding: Defending against Jailbreak Attacks via Safety-Aware Decoding. (38%)Zhangchen Xu; Fengqing Jiang; Luyao Niu; Jinyuan Jia; Bill Yuchen Lin; Radha Poovendran Rapid Adoption, Hidden Risks: The Dual Impact of Large Language Model Customization. (9%)Rui Zhang; Hongwei Li; Rui Wen; Wenbo Jiang; Yuan Zhang; Michael Backes; Yun Shen; Yang Zhang Reward Poisoning Attack Against Offline Reinforcement Learning. (5%)Yinglun Xu; Rohan Gumaste; Gagandeep Singh Adversarial Nibbler: An Open Red-Teaming Method for Identifying Diverse Harms in Text-to-Image Generation. (3%)Jessica Quaye; Alicia Parrish; Oana Inel; Charvi Rastogi; Hannah Rose Kirk; Minsuk Kahng; Liemt Erin van; Max Bartolo; Jess Tsang; Justin White; Nathan Clement; Rafael Mosquera; Juan Ciro; Vijay Janapa Reddi; Lora Aroyo Ten Words Only Still Help: Improving Black-Box AI-Generated Text Detection via Proxy-Guided Efficient Re-Sampling. (2%)Yuhui Shi; Qiang Sheng; Juan Cao; Hao Mi; Beizhe Hu; Danding Wang Towards Robust Model-Based Reinforcement Learning Against Adversarial Corruption. (1%)Chenlu Ye; Jiafan He; Quanquan Gu; Tong Zhang Immediate generalisation in humans but a generalisation lag in deep neural networks$\unicode{x2014}$evidence for representational divergence? (1%)Lukas S. Huber; Fred W. Mast; Felix A. Wichmann Play Guessing Game with LLM: Indirect Jailbreak Attack with Implicit Clues. (1%)Zhiyuan Chang; Mingyang Li; Yi Liu; Junjie Wang; Qing Wang; Yang Liu 2024-02-13 Faster Repeated Evasion Attacks in Tree Ensembles. (96%)Lorenzo Cascioli; Laurens Devos; Ondřej Kuželka; Jesse Davis Generating Universal Adversarial Perturbations for Quantum Classifiers. (93%)Gautham Anil; Vishnu Vinod; Apurva Narayan Enhancing Robustness of Indoor Robotic Navigation with Free-Space Segmentation Models Against Adversarial Attacks. (83%)Qiyuan An; Christos Sevastopoulos; Fillia Makedon Data Reconstruction Attacks and Defenses: A Systematic Evaluation. (76%)Sheng Liu; Zihan Wang; Qi Lei COLD-Attack: Jailbreaking LLMs with Stealthiness and Controllability. (62%)Xingang Guo; Fangxu Yu; Huan Zhang; Lianhui Qin; Bin Hu Test-Time Backdoor Attacks on Multimodal Large Language Models. (56%)Dong Lu; Tianyu Pang; Chao Du; Qian Liu; Xianjun Yang; Min Lin Adversarially Robust Feature Learning for Breast Cancer Diagnosis. (33%)Degan Hao; Dooman Arefan; Margarita Zuley; Wendie Berg; Shandong Wu Agent Smith: A Single Image Can Jailbreak One Million Multimodal LLM Agents Exponentially Fast. (31%)Xiangming Gu; Xiaosen Zheng; Tianyu Pang; Chao Du; Qian Liu; Ye Wang; Jing Jiang; Min Lin Feature Attribution with Necessity and Sufficiency via Dual-stage Perturbation Test for Causal Explanation. (1%)Xuexin Chen; Ruichu Cai; Zhengting Huang; Yuxuan Zhu; Julien Horwood; Zhifeng Hao; Zijian Li; Jose Miguel Hernandez-Lobato 2024-02-12 Understanding Deep Learning defenses Against Adversarial Examples Through Visualizations for Dynamic Risk Assessment. (99%)Xabier Echeberria-Barrio; Amaia Gil-Lerchundi; Jon Egana-Zubia; Raul Orduna-Urrutia Topological safeguard for evasion attack interpreting the neural networks' behavior. (89%)Xabier Echeberria-Barrio; Amaia Gil-Lerchundi; Iñigo Mendialdua; Raul Orduna-Urrutia PoisonedRAG: Knowledge Poisoning Attacks to Retrieval-Augmented Generation of Large Language Models. (83%)Wei Zou; Runpeng Geng; Binghui Wang; Jinyuan Jia Privacy-Preserving Gaze Data Streaming in Immersive Interactive Virtual Reality: Robustness and User Experience. (33%)Ethan Wilson; Azim Ibragimov; Michael J. Proulx; Sai Deep Tetali; Kevin Butler; Eakta Jain OrderBkd: Textual backdoor attack through repositioning. (13%)Irina Alekseevskaia; Konstantin Arkhipenko Tighter Bounds on the Information Bottleneck with Application to Deep Learning. (10%)Nir Weingarten; Zohar Yakhini; Moshe Butman; Ran Gilad-Bachrach Customizable Perturbation Synthesis for Robust SLAM Benchmarking. (9%)Xiaohao Xu; Tianyi Zhang; Sibo Wang; Xiang Li; Yongqi Chen; Ye Li; Bhiksha Raj; Matthew Johnson-Roberson; Xiaonan Huang Multi-Attribute Vision Transformers are Efficient and Robust Learners. (8%)Hanan Gani; Nada Saadi; Noor Hussein; Karthik Nandakumar Accelerated Smoothing: A Scalable Approach to Randomized Smoothing. (3%)Devansh Bhardwaj; Kshitiz Kaushik; Sarthak Gupta Game of Trojans: Adaptive Adversaries Against Output-based Trojaned-Model Detectors. (3%)Dinuka Sahabandu; Xiaojun Xu; Arezoo Rajabi; Luyao Niu; Bhaskar Ramasubramanian; Bo Li; Radha Poovendran THE COLOSSEUM: A Benchmark for Evaluating Generalization for Robotic Manipulation. (2%)Wilbert Pumacay; Ishika Singh; Jiafei Duan; Ranjay Krishna; Jesse Thomason; Dieter Fox Local Centrality Minimization with Quality Guarantees. (1%)Atsushi Miyauchi; Lorenzo Severini; Francesco Bonchi Do Membership Inference Attacks Work on Large Language Models? (1%)Michael Duan; Anshuman Suri; Niloofar Mireshghallah; Sewon Min; Weijia Shi; Luke Zettlemoyer; Yulia Tsvetkov; Yejin Choi; David Evans; Hannaneh Hajishirzi NeuralSentinel: Safeguarding Neural Network Reliability and Trustworthiness. (1%)Xabier Echeberria-Barrio; Mikel Gorricho; Selene Valencia; Francesco Zola Pixel Sentence Representation Learning. (1%)Chenghao Xiao; Zhuoxu Huang; Danlu Chen; G Thomas Hudson; Yizhi Li; Haoran Duan; Chenghua Lin; Jie Fu; Jungong Han; Noura Al Moubayed 2024-02-11 A Random Ensemble of Encrypted Vision Transformers for Adversarially Robust Defense. (99%)Ryota Iijima; Sayaka Shiota; Hitoshi Kiya Accuracy of TextFooler black box adversarial attacks on 01 loss sign activation neural network ensemble. (98%)Yunzhe Xue; Usman Roshan 2024-02-10 Whispers in the Machine: Confidentiality in LLM-integrated Systems. (26%)Jonathan Evertz; Merlin Chlosta; Lea Schönherr; Thorsten Eisenhofer Architectural Neural Backdoors from First Principles. (26%)Harry Langford; Ilia Shumailov; Yiren Zhao; Robert Mullins; Nicolas Papernot 2024-02-09 Anomaly Unveiled: Securing Image Classification against Adversarial Patch Attacks. (98%)Nandish Chattopadhyay; Amira Guesmi; Muhammad Shafique Studious Bob Fight Back Against Jailbreaking via Prompt Adversarial Tuning. (95%)Yichuan Mo; Yuji Wang; Zeming Wei; Yisen Wang RAMP: Boosting Adversarial Robustness Against Multiple $l_p$ Perturbations. (83%)Enyi Jiang; Gagandeep Singh System-level Analysis of Adversarial Attacks and Defenses on Intelligence in O-RAN based Cellular Networks. (82%)Azuka Chiejina; Brian Kim; Kaushik Chowhdury; Vijay K. Shah The SpongeNet Attack: Sponge Weight Poisoning of Deep Neural Networks. (69%)Jona te Lintelo; Stefanos Koffas; Stjepan Picek Corruption Robust Offline Reinforcement Learning with Human Feedback. (67%)Debmalya Mandal; Andi Nika; Parameswaran Kamalaruban; Adish Singla; Goran Radanović Quantifying and Enhancing Multi-modal Robustness with Modality Preference. (56%)Zequn Yang; Yake Wei; Ce Liang; Di Hu StruQ: Defending Against Prompt Injection with Structured Queries. (45%)Sizhe Chen; Julien Piet; Chawin Sitawarin; David Wagner Evaluating Membership Inference Attacks and Defenses in Federated Learning. (4%)Gongxi Zhu; Donghao Li; Hanlin Gu; Yuxing Han; Yuan Yao; Lixin Fan; Qiang Yang For Better or For Worse? Learning Minimum Variance Features With Label Augmentation. (1%)Muthu Chidambaram; Rong Ge 2024-02-08 Comprehensive Assessment of Jailbreak Attacks Against LLMs. (99%)Junjie Chu; Yugeng Liu; Ziqing Yang; Xinyue Shen; Michael Backes; Yang Zhang Investigating White-Box Attacks for On-Device Models. (93%)Mingyi Zhou; Xiang Gao; Jing Wu; Kui Liu; Hailong Sun; Li Li TETRIS: Towards Exploring the Robustness of Interactive Segmentation. (81%)Andrey Moskalenko; Vlad Shakhuro; Anna Vorontsova; Anton Konushin; Anton Antonov; Alexander Krapukhin; Denis Shepelev; Konstantin Soshin Linearizing Models for Efficient yet Robust Private Inference. (68%)Sreetama Sarkar; Souvik Kundu; Peter A. Beerel A High Dimensional Model for Adversarial Training: Geometry and Trade-Offs. (26%)Kasimir Tanner; Matteo Vilucchio; Bruno Loureiro; Florent Krzakala Is Adversarial Training with Compressed Datasets Effective? (10%)Tong Chen; Raghavendra Selvan Reinforcement Learning as a Catalyst for Robust and Fair Federated Learning: Deciphering the Dynamics of Client Contributions. (9%)Jialuo He; Wei Chen; Xiaojin Zhang 2024-02-07 Adversarial Robustness Through Artifact Design. (99%)Tsufit Shua; Mahmood Sharif EvoSeed: Unveiling the Threat on Deep Neural Networks with Real-World Illusions. (98%)Shashank Kotyan; PoYuan Mao; Danilo Vasconcellos Vargas Analyzing Adversarial Inputs in Deep Reinforcement Learning. (96%)Davide Corsi; Guy Amir; Guy Katz; Alessandro Farinelli Assessing the Brittleness of Safety Alignment via Pruning and Low-Rank Modifications. (1%)Boyi Wei; Kaixuan Huang; Yangsibo Huang; Tinghao Xie; Xiangyu Qi; Mengzhou Xia; Prateek Mittal; Mengdi Wang; Peter Henderson 2024-02-06 Boosting Adversarial Transferability across Model Genus by Deformation-Constrained Warping. (98%)Qinliang Lin; Cheng Luo; Zenghao Niu; Xilin He; Weicheng Xie; Yuanbo Hou; Linlin Shen; Siyang Song Adversarially Robust Deepfake Detection via Adversarial Feature Similarity Learning. (98%)Sarwar Khan PAC-Bayesian Adversarially Robust Generalization Bounds for Graph Neural Network. (75%)Tan Sun; Junhong Lin Enhance DNN Adversarial Robustness and Efficiency via Injecting Noise to Non-Essential Neurons. (74%)Zhenyu Liu; Garrett Gagnon; Swagath Venkataramani; Liu Liu SUB-PLAY: Adversarial Policies against Partially Observed Multi-Agent Reinforcement Learning Systems. (67%)Oubo Ma; Yuwen Pu; Linkang Du; Yang Dai; Ruo Wang; Xiaolei Liu; Yingcai Wu; Shouling Ji BotSSCL: Social Bot Detection with Self-Supervised Contrastive Learning. (64%)Mohammad Majid Akhtar; Navid Shadman Bhuiyan; Rahat Masood; Muhammad Ikram; Salil S. Kanhere Studying Vulnerable Code Entities in R. (10%)Zixiao Zhao; Millon Madhur Das; Fatemeh H. Fard DeMarking: A Defense for Network Flow Watermarking in Real-Time. (10%)Yali Yuan; Jian Ge; Guang Cheng HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusal. (2%)Mantas Mazeika; Long Phan; Xuwang Yin; Andy Zou; Zifan Wang; Norman Mu; Elham Sakhaee; Nathaniel Li; Steven Basart; Bo Li; David Forsyth; Dan Hendrycks 2024-02-05 A Generative Approach to Surrogate-based Black-box Attacks. (99%)Raha Moraffah; Huan Liu Transcending Adversarial Perturbations: Manifold-Aided Adversarial Examples with Legitimate Semantics. (99%)Shuai Li; Xiaoyu Jiang; Xiaoguang Ma Arabic Synonym BERT-based Adversarial Examples for Text Classification. (99%)Norah Alshahrani; Saied Alshahrani; Esma Wali; Jeanna Matthews Generalization Properties of Adversarial Training for $\ell_0$-Bounded Adversarial Attacks. (92%)Payam Delgosha; Hamed Hassani; Ramtin Pedarsani FoolSDEdit: Deceptively Steering Your Edits Towards Targeted Attribute-aware Distribution. (89%)Qi Zhou; Dongxia Wang; Tianlin Li; Zhihong Xu; Yang Liu; Kui Ren; Wenhai Wang; Qing Guo Time-Distributed Backdoor Attacks on Federated Spiking Learning. (83%)Gorka Abad; Stjepan Picek; Aitor Urbieta Shadowcast: Stealthy Data Poisoning Attacks Against Vision-Language Models. (83%)Yuancheng Xu; Jiarui Yao; Manli Shu; Yanchao Sun; Zichu Wu; Ning Yu; Tom Goldstein; Furong Huang Partially Recentralization Softmax Loss for Vision-Language Models Robustness. (81%)Hao Wang; Xin Zhang; Jinzhe Jiang; Yaqian Zhao; Chen Li Organic or Diffused: Can We Distinguish Human Art from AI-generated Images? (31%)Anna Yoo Jeong Ha; Josephine Passananti; Ronik Bhaskar; Shawn Shan; Reid Southen; Haitao Zheng; Ben Y. Zhao DisDet: Exploring Detectability of Backdoor Attack on Diffusion Models. (12%)Yang Sui; Huy Phan; Jinqi Xiao; Tianfang Zhang; Zijie Tang; Cong Shi; Yan Wang; Yingying Chen; Bo Yuan FINEST: Stabilizing Recommendations by Rank-Preserving Fine-Tuning. (1%)Sejoon Oh; Berk Ustun; Julian McAuley; Srijan Kumar 2024-02-04 PROSAC: Provably Safe Certification for Machine Learning Models under Adversarial Attacks. (99%)Ziquan Liu; Zhuo Zhi; Ilija Bogunovic; Carsten Gerner-Beuerle; Miguel Rodrigues DeSparsify: Adversarial Attack Against Token Sparsification Mechanisms in Vision Transformers. (99%)Oryan Yehezkel; Alon Zolfi; Amit Baras; Yuval Elovici; Asaf Shabtai Adversarial Text Purification: A Large Language Model Approach for Defense. (99%)Raha Moraffah; Shubh Khandelwal; Amrita Bhattacharjee; Huan Liu Exploiting Class Probabilities for Black-box Sentence-level Attacks. (75%)Raha Moraffah; Huan Liu Evading Deep Learning-Based Malware Detectors via Obfuscation: A Deep Reinforcement Learning Approach. (41%)Brian Etter; James Lee Hu; Mohammedreza Ebrahimi; Weifeng Li; Xin Li; Hsinchun Chen Adversarial Data Augmentation for Robust Speaker Verification. (1%)Zhenyu Zhou; Junhui Chen; Namin Wang; Lantian Li; Dong Wang 2024-02-03 Evaluating the Robustness of Off-Road Autonomous Driving Segmentation against Adversarial Attacks: A Dataset-Centric analysis. (96%)Pankaj Deoli; Rohit Kumar; Axel Vierling; Karsten Berns Your Diffusion Model is Secretly a Certifiably Robust Classifier. (80%)Huanran Chen; Yinpeng Dong; Shitong Shao; Zhongkai Hao; Xiao Yang; Hang Su; Jun Zhu MixedNUTS: Training-Free Accuracy-Robustness Balance via Nonlinearly Mixed Classifiers. (76%)Yatong Bai; Mo Zhou; Vishal M. Patel; Somayeh Sojoudi Analyzing Sentiment Polarity Reduction in News Presentation through Contextual Perturbation and Large Language Models. (68%)Alapan Kuila; Somnath Jena; Sudeshna Sarkar; Partha Pratim Chakrabarti Universal Post-Training Reverse-Engineering Defense Against Backdoors in Deep Neural Networks. (16%)Xi Li; Hang Wang; David J. Miller; George Kesidis Towards Optimal Adversarial Robust Q-learning with Bellman Infinity-error. (10%)Haoran Li; Zicheng Zhang; Wang Luo; Congying Han; Yudong Hu; Tiande Guo; Shichen Liao Invisible Finger: Practical Electromagnetic Interference Attack on Touchscreen-based Electronic Devices. (9%)Haoqi Shan; Boyi Zhang; Zihao Zhan; Dean Sullivan; Shuo Wang; Yier Jin Safety Fine-Tuning at (Almost) No Cost: A Baseline for Vision Large Language Models. (5%)Yongshuo Zong; Ondrej Bohdal; Tingyang Yu; Yongxin Yang; Timothy Hospedales Data Poisoning for In-context Learning. (5%)Pengfei He; Han Xu; Yue Xing; Hui Liu; Makoto Yamada; Jiliang Tang Seeing is not always believing: The Space of Harmless Perturbations. (2%)Lu Chen; Shaofeng Li; Benhao Huang; Fan Yang; Zheng Li; Jie Li; Yuan Luo 2024-02-02 $\sigma$-zero: Gradient-based Optimization of $\ell_0$-norm Adversarial Examples. (99%)Antonio Emanuele Cinà; Francesco Villani; Maura Pintor; Lea Schönherr; Battista Biggio; Marcello Pelillo HQA-Attack: Toward High Quality Black-Box Hard-Label Adversarial Attack on Text. (99%)Han Liu; Zhi Xu; Xiaotong Zhang; Feng Zhang; Fenglong Ma; Hongyang Chen; Hong Yu; Xianchao Zhang STAA-Net: A Sparse and Transferable Adversarial Attack for Speech Emotion Recognition. (99%)Yi Chang; Zhao Ren; Zixing Zhang; Xin Jing; Kun Qian; Xi Shao; Bin Hu; Tanja Schultz; Björn W. Schuller Delving into Decision-based Black-box Attacks on Semantic Segmentation. (93%)Zhaoyu Chen; Zhengyang Shan; Jingwen Chang; Kaixun Jiang; Dingkang Yang; Yiting Cheng; Wenqiang Zhang SignSGD with Federated Defense: Harnessing Adversarial Attacks through Gradient Sign Decoding. (92%)Chanho Park; Namyoon Lee Unlearnable Examples For Time Series. (86%)Yujing Jiang; Xingjun Ma; Sarah Monazam Erfani; James Bailey Preference Poisoning Attacks on Reward Model Learning. (83%)Junlin Wu; Jiongxiao Wang; Chaowei Xiao; Chenguang Wang; Ning Zhang; Yevgeniy Vorobeychik S2malloc: Statistically Secure Allocator for Use-After-Free Protection And More. (3%)Ruizhe Wang; Meng Xu; N. Asokan Privacy-Preserving Distributed Learning for Residential Short-Term Load Forecasting. (3%)Yi Dong; Yingjie Wang; Mariana Gama; Mustafa A. Mustafa; Geert Deconinck; Xiaowei Huang Cheating Suffix: Targeted Attack to Text-To-Image Diffusion Models with Multi-Modal Priors. (2%)Dingcheng Yang; Yang Bai; Xiaojun Jia; Yang Liu; Xiaochun Cao; Wenjian Yu What Will My Model Forget? Forecasting Forgotten Examples in Language Model Refinement. (1%)Xisen Jin; Xiang Ren 2024-02-01 Benchmarking Transferable Adversarial Attacks. (98%)Zhibo Jin; Jiayu Zhang; Zhiyu Zhu; Huaming Chen Hidding the Ghostwriters: An Adversarial Evaluation of AI-Generated Student Essay Detection. (70%)Xinlin Peng; Ying Zhou; Ben He; Le Sun; Yingfei Sun Double-Dip: Thwarting Label-Only Membership Inference Attacks with Transfer Learning and Randomization. (64%)Arezoo Rajabi; Reeya Pimple; Aiswarya Janardhanan; Surudhi Asokraj; Bhaskar Ramasubramanian; Radha Poovendran Vision-LLMs Can Fool Themselves with Self-Generated Typographic Attacks. (45%)Maan Qraitem; Nazia Tasnim; Piotr Teterwak; Kate Saenko; Bryan A. Plummer Approximating Optimal Morphing Attacks using Template Inversion. (9%)Laurent Colbois; Hatef Otroshi Shahreza; Sébastien Marcel Trustworthy Distributed AI Systems: Robustness, Privacy, and Governance. (8%)Wenqi Wei; Ling Liu algoXSSF: Detection and analysis of cross-site request forgery (XSRF) and cross-site scripting (XSS) attacks via Machine learning algorithms. (1%)Naresh Kshetri; Dilip Kumar; James Hutson; Navneet Kaur; Omar Faruq Osama 2024-01-31 Adversarial Quantum Machine Learning: An Information-Theoretic Generalization Analysis. (95%)Petros Georgiou; Sharu Theresa Jose; Osvaldo Simeone Invariance-powered Trustworthy Defense via Remove Then Restore. (70%)Xiaowei Fu; Yuhang Zhou; Lina Ma; Lei Zhang BrainLeaks: On the Privacy-Preserving Properties of Neuromorphic Architectures against Model Inversion Attacks. (13%)Hamed Poursiami; Ihsen Alouani; Maryam Parsa LoRec: Large Language Model for Robust Sequential Recommendation against Poisoning Attacks. (9%)Kaike Zhang; Qi Cao; Yunfan Wu; Fei Sun; Huawei Shen; Xueqi Cheng Logit Poisoning Attack in Distillation-based Federated Learning and its Countermeasures. (4%)Yonghao Yu; Shunan Zhu; Jinglu Hu Manipulating Predictions over Discrete Inputs in Machine Teaching. (1%)Xiaodong Wu; Yufei Han; Hayssam Dahrouj; Jianbing Ni; Zhenwen Liang; Xiangliang Zhang Ambush from All Sides: Understanding Security Threats in Open-Source Software CI/CD Pipelines. (1%)Ziyue Pan; Wenbo Shen; Xingkai Wang; Yutian Yang; Rui Chang; Yao Liu; Chengwei Liu; Yang Liu; Kui Ren 2024-01-30 Single Word Change is All You Need: Designing Attacks and Defenses for Text Classifiers. (99%)Lei Xu; Sarah Alnegheimish; Laure Berti-Equille; Alfredo Cuesta-Infante; Kalyan Veeramachaneni Robust Prompt Optimization for Defending Language Models Against Jailbreaking Attacks. (98%)Andy Zhou; Bo Li; Haohan Wang Towards Assessing the Synthetic-to-Measured Adversarial Vulnerability of SAR ATR. (98%)Bowen Peng; Bo Peng; Jingyuan Xia; Tianpeng Liu; Yongxiang Liu; Li Liu AdvGPS: Adversarial GPS for Multi-Agent Perception Attack. (95%)Jinlong Li; Baolu Li; Xinyu Liu; Jianwu Fang; Felix Juefei-Xu; Qing Guo; Hongkai Yu Game-Theoretic Unlearnable Example Generator. (92%)Shuang Liu; Yihan Wang; Xiao-Shan Gao Camouflage Adversarial Attacks on Multiple Agent Systems. (87%)Ziqing Lu; Guanlin Liu; Lifeng Lai; Weiyu Xu Weak-to-Strong Jailbreaking on Large Language Models. (76%)Xuandong Zhao; Xianjun Yang; Tianyu Pang; Chao Du; Lei Li; Yu-Xiang Wang; William Yang Wang A Proactive and Dual Prevention Mechanism against Illegal Song Covers empowered by Singing Voice Conversion. (75%)Guangke Chen; Yedi Zhang; Fu Song; Ting Wang; Xiaoning Du; Yang Liu Improving QA Model Performance with Cartographic Inoculation. (26%)Allen UT Austin Chen; Okan UT Austin Tanrikulu Towards Visual Syntactical Understanding. (4%)Sayeed Shafayet Chowdhury; Soumyadeep Chandra; Kaushik Roy Provably Robust Multi-bit Watermarking for AI-generated Text via Error Correction Code. (2%)Wenjie Qu; Dong Yin; Zixin He; Wei Zou; Tianyang Tao; Jinyuan Jia; Jiaheng Zhang 2024-01-29 LESSON: Multi-Label Adversarial False Data Injection Attack for Deep Learning Locational Detection. (99%)Jiwei Tian; Chao Shen; Buhong Wang; Xiaofang Xia; Meng Zhang; Chenhao Lin; Qian Li Adversarial Training on Purification (AToP): Advancing Both Robustness and Generalization. (92%)Guang Lin; Chao Li; Jianhai Zhang; Toshihisa Tanaka; Qibin Zhao Revisiting Gradient Pruning: A Dual Realization for Defending against Gradient Attacks. (68%)Lulu Xue; Shengshan Hu; Ruizhi Zhao; Leo Yu Zhang; Shengqing Hu; Lichao Sun; Dezhong Yao GPS: Graph Contrastive Learning via Multi-scale Augmented Views from Adversarial Pooling. (5%)Wei Ju; Yiyang Gu; Zhengyang Mao; Ziyue Qiao; Yifang Qin; Xiao Luo; Hui Xiong; Ming Zhang Security and Privacy Challenges of Large Language Models: A Survey. (1%)Badhan Chandra Das; M. Hadi Amini; Yanzhao Wu 2024-01-28 Addressing Noise and Efficiency Issues in Graph-Based Machine Learning Models From the Perspective of Adversarial Attack. (81%)Yongyu Wang Transparency Attacks: How Imperceptible Image Layers Can Fool AI Perception. (75%)Forrest McKee; David Noever 2024-01-27 L-AutoDA: Leveraging Large Language Models for Automated Decision-based Adversarial Attacks. (98%)Ping Guo; Fei Liu; Xi Lin; Qingchuan Zhao; Qingfu Zhang 2024-01-26 End-To-End Set-Based Training for Neural Network Verification. (99%)Lukas Koller; Tobias Ladner; Matthias Althoff Mitigating Feature Gap for Adversarial Robustness by Feature Disentanglement. (91%)Nuoyan Zhou; Dawei Zhou; Decheng Liu; Xinbo Gao; Nannan Wang Multi-Trigger Backdoor Attacks: More Triggers, More Threats. (82%)Yige Li; Xingjun Ma; Jiabo He; Hanxun Huang; Yu-Gang Jiang Adversarial Attacks and Defenses in 6G Network-Assisted IoT Systems. (81%)Bui Duc Son; Nguyen Tien Hoa; Chien Trinh Van; Waqas Khalid; Mohamed Amine Ferrag; Wan Choi; Merouane Debbah Conserve-Update-Revise to Cure Generalization and Robustness Trade-off in Adversarial Training. (62%)Shruthi Gowda; Bahram Zonooz; Elahe Arani Asymptotic Behavior of Adversarial Training Estimator under $\ell_\infty$-Perturbation. (22%)Yiling Xie; Xiaoming Huo Better Representations via Adversarial Training in Pre-Training: A Theoretical Perspective. (22%)Yue Xing; Xiaofeng Lin; Qifan Song; Yi Xu; Belinda Zeng; Guang Cheng MEA-Defender: A Robust Watermark against Model Extraction Attack. (13%)Peizhuo Lv; Hualong Ma; Kai Chen; Jiachen Zhou; Shengzhi Zhang; Ruigang Liang; Shenchen Zhu; Pan Li; Yingjun Zhang BackdoorBench: A Comprehensive Benchmark and Analysis of Backdoor Learning. (2%)Baoyuan Wu; Hongrui Chen; Mingda Zhang; Zihao Zhu; Shaokui Wei; Danni Yuan; Mingli Zhu; Ruotong Wang; Li Liu; Chao Shen 2024-01-25 Sparse and Transferable Universal Singular Vectors Attack. (99%)Kseniia Kuvshinova; Olga Tsymboi; Ivan Oseledets Friendly Attacks to Improve Channel Coding Reliability. (54%)Anastasiia Kurmukova; Deniz Gunduz Semantic Sensitivities and Inconsistent Predictions: Measuring the Fragility of NLI Models. (16%)Erik Arakelyan; Zhaoqi Liu; Isabelle Augenstein The Risk of Federated Learning to Skew Fine-Tuning Features and Underperform Out-of-Distribution Robustness. (2%)Mengyao Du; Miao Zhang; Yuwen Pu; Kai Xu; Shouling Ji; Quanjun Yin Novel Quadratic Constraints for Extending LipSDP beyond Slope-Restricted Activations. (1%)Patricia Pauli; Aaron Havens; Alexandre Araujo; Siddharth Garg; Farshad Khorrami; Frank Allgöwer; Bin Hu Physical Trajectory Inference Attack and Defense in Decentralized POI Recommendation. (1%)Jing Long; Tong Chen; Guanhua Ye; Kai Zheng; Nguyen Quoc Viet Hung; Hongzhi Yin 2024-01-24 Can overfitted deep neural networks in adversarial training generalize? -- An approximation viewpoint. (86%)Zhongjie Shi; Fanghui Liu; Yuan Cao; Johan A. K. Suykens A Systematic Approach to Robustness Modelling for Deep Convolutional Neural Networks. (83%)Charles Meyers; Mohammad Reza Saleh Sedghpour; Tommy Löfstedt; Erik Elmroth WPDA: Frequency-based Backdoor Attack with Wavelet Packet Decomposition. (76%)Zhengyao Song; Yongqiang Li; Danni Yuan; Li Liu; Shaokui Wei; Baoyuan Wu Exploring Adversarial Threat Models in Cyber Physical Battery Systems. (76%)Shanthan Kumar Padisala; Shashank Dhananjay Vyas; Satadru Dey Fluent dreaming for language models. (64%)T. Ben Confirm Labs Thompson; Zygimantas Confirm Labs Straznickas; Michael Confirm Labs Sklar 2024-01-23 Boosting the Transferability of Adversarial Examples via Local Mixup and Adaptive Step Size. (99%)Junlin Liu; Xinchen Lyu Securing Recommender System via Cooperative Training. (80%)Qingyang Wang; Chenwang Wu; Defu Lian; Enhong Chen Compositional Generative Inverse Design. (56%)Tailin Wu; Takashi Maruyama; Long Wei; Tao Zhang; Yilun Du; Gianluca Iaccarino; Jure Leskovec AdCorDA: Classifier Refinement via Adversarial Correction and Domain Adaptation. (33%)Lulan Shen; Ali Edalati; Brett Meyer; Warren Gross; James J. Clark ToDA: Target-oriented Diffusion Attacker against Recommendation System. (13%)Xiaohao Liu; Zhulin Tao; Ting Jiang; He Chang; Yunshan Ma; Xianglin Huang; Xiang Wang DAFA: Distance-Aware Fair Adversarial Training. (2%)Hyungyu Lee; Saehyung Lee; Hyemi Jang; Junsung Park; Ho Bae; Sungroh Yoon The twin peaks of learning neural networks. (2%)Elizaveta Demyanenko; Christoph Feinauer; Enrico M. Malatesta; Luca Saglietti 2024-01-22 Fast Adversarial Training against Textual Adversarial Attacks. (99%)Yichen Yang; Xin Liu; Kun He A Training-Free Defense Framework for Robust Learned Image Compression. (74%)Myungseo Song; Jinyoung Choi; Bohyung Han Adversarial speech for voice privacy protection from Personalized Speech generation. (73%)Shihao Chen; Liping Chen; Jie Zhang; KongAik Lee; Zhenhua Ling; Lirong Dai NEUROSEC: FPGA-Based Neuromorphic Audio Security. (13%)Murat Isik; Hiruna Vishwamith; Yusuf Sur; Kayode Inadagbo; I. Can Dikmen Unraveling Attacks in Machine Learning-based IoT Ecosystems: A Survey and the Open Libraries Behind Them. (13%)Chao Liu; Boxi Chen; Wei Shao; Chris Zhang; Kelvin Wong; Yi Zhang Robustness to distribution shifts of compressed networks for edge devices. (8%)Lulan Shen; Ali Edalati; Brett Meyer; Warren Gross; James J. Clark Text Embedding Inversion Security for Multilingual Language Models. (2%)Yiyi Chen; Heather Lent; Johannes Bjerva Out-of-Distribution Detection & Applications With Ablated Learned Temperature Energy. (1%)Will LeVine; Benjamin Pikus; Jacob Phillips; Berk Norman; Fernando Amat Gil; Sean Hendryx 2024-01-21 How Robust Are Energy-Based Models Trained With Equilibrium Propagation? (99%)Siddharth Mansingh; Michal Kucer; Garrett Kenyon; Juston Moore; Michael Teti Analyzing the Quality Attributes of AI Vision Models in Open Repositories Under Adversarial Attacks. (56%)Zerui Wang; Yan Liu Adversarial Augmentation Training Makes Action Recognition Models More Robust to Realistic Video Distribution Shifts. (11%)Kiyoon Kim; Shreyank N Gowda; Panagiotis Eustratiadis; Antreas Antoniou; Robert B Fisher Efficient local linearity regularization to overcome catastrophic overfitting. (8%)Elias Abad Rocamora; Fanghui Liu; Grigorios G. Chrysos; Pablo M. Olmos; Volkan Cevher 2024-01-20 Susceptibility of Adversarial Attack on Medical Image Segmentation Models. (99%)Zhongxuan Wang; Leo Xu Finding a Needle in the Adversarial Haystack: A Targeted Paraphrasing Approach For Uncovering Edge Cases with Minimal Distribution Distortion. (96%)Aly M. Kassem; Sherif Saad CARE: Ensemble Adversarial Robustness Evaluation Against Adaptive Attackers for Security Applications. (80%)Hangsheng Zhang; Jiqiang Liu; Jinsong Dong Inducing High Energy-Latency of Large Vision-Language Models with Verbose Images. (33%)Kuofeng Gao; Yang Bai; Jindong Gu; Shu-Tao Xia; Philip Torr; Zhifeng Li; Wei Liu 2024-01-19 PuriDefense: Randomized Local Implicit Adversarial Purification for Defending Black-box Query-based Attacks. (99%)Ping Guo; Zhiyuan Yang; Xi Lin; Qingchuan Zhao; Qingfu Zhang Explainable and Transferable Adversarial Attack for ML-Based Network Intrusion Detectors. (99%)Hangsheng Zhang; Dongqi Han; Yinlong Liu; Zhiliang Wang; Jiyan Sun; Shangyuan Zhuang; Jiqiang Liu; Jinsong Dong The Surprising Harmfulness of Benign Overfitting for Adversarial Robustness. (98%)Yifan Hao; Tong Zhang FIMBA: Evaluating the Robustness of AI in Genomics via Feature Importance Adversarial Attacks. (56%)Heorhii Skovorodnikov; Hoda Alkhzaimi Adversarially Robust Signed Graph Contrastive Learning from Balance Augmentation. (10%)Jialong Zhou; Xing Ai; Yuni Lai; Kai Zhou BadChain: Backdoor Chain-of-Thought Prompting for Large Language Models. (3%)Zhen Xiang; Fengqing Jiang; Zidi Xiong; Bhaskar Ramasubramanian; Radha Poovendran; Bo Li Image Safeguarding: Reasoning with Conditional Vision Language Model and Obfuscating Unsafe Content Counterfactually. (1%)Mazal Bethany; Brandon Wherry; Nishant Vishwamitra; Peyman Najafirad 2024-01-18 HGAttack: Transferable Heterogeneous Graph Adversarial Attack. (99%)He Zhao; Zhiwei Zeng; Yongwei Wang; Deheng Ye; Chunyan Miao Hijacking Attacks against Neural Networks by Analyzing Training Data. (99%)Yunjie Ge; Qian Wang; Huayang Huang; Qi Li; Cong Wang; Chao Shen; Lingchen Zhao; Peipei Jiang; Zheng Fang; Shenyi Zhang Marrying Adapters and Mixup to Efficiently Enhance the Adversarial Robustness of Pre-Trained Language Models for Text Classification. (98%)Tuc Nguyen; Thai Le Hacking Predictors Means Hacking Cars: Using Sensitivity Analysis to Identify Trajectory Prediction Vulnerabilities for Autonomous Driving Security. (92%)Marsalis Gibson; David Babazadeh; Claire Tomlin; Shankar Sastry Differentially Private and Adversarially Robust Machine Learning: An Empirical Evaluation. (80%)Janvi Thakkar; Giulio Zizzo; Sergio Maffeis Investigating Training Strategies and Model Robustness of Low-Rank Adaptation for Language Modeling in Speech Recognition. (15%)Yu Yu; Chao-Han Huck Yang; Tuan Dinh; Sungho Ryu; Jari Kolehmainen; Roger Ren; Denis Filimonov; Prashanth G. Shivakumar; Ankur Gandhe; Ariya Rastow; Jia Xu; Ivan Bulyko; Andreas Stolcke Power in Numbers: Robust reading comprehension by finetuning with four adversarial sentences per example. (13%)Ariel Marcus Cross-Modality Perturbation Synergy Attack for Person Re-identification. (2%)Yunpeng Gong; Zhun Zhong; Zhiming Luo; Yansong Qu; Rongrong Ji; Min Jiang Vulnerabilities of Foundation Model Integrated Federated Learning Under Adversarial Threats. (2%)Chen Wu; Xi Li; Jiaqi Wang Large Language Models are Efficient Learners of Noise-Robust Speech Recognition. (1%)Yuchen Hu; Chen Chen; Chao-Han Huck Yang; Ruizhe Li; Chao Zhang; Pin-Yu Chen; EnSiong Chng 2024-01-17 Towards Scalable and Robust Model Versioning. (93%)Wenxin Ding; Arjun Nitin Bhagoji; Ben Y. Zhao; Haitao Zheng Artwork Protection Against Neural Style Transfer Using Locally Adaptive Adversarial Color Attack. (93%)Zhongliang Guo; Kaixuan Wang; Weiye Li; Yifei Qian; Ognjen Arandjelović; Lei Fang A GAN-based data poisoning framework against anomaly detection in vertical federated learning. (3%)Xiaolin Chen; Daoguang Zan; Wei Li; Bei Guan; Yongji Wang An Optimal Transport Approach for Computing Adversarial Training Lower Bounds in Multiclass Classification. (3%)Nicolas Garcia Trillos; Matt Jacobs; Jakwang Kim; Matthew Werenski MITS-GAN: Safeguarding Medical Imaging from Tampering with Generative Adversarial Networks. (2%)Giovanni Pasqualino; Luca Guarnera; Alessandro Ortis; Sebastiano Battiato Caught in the Quicksand of Reasoning, Far from AGI Summit: Evaluating LLMs' Mathematical and Coding Competency through Ontology-guided Interventions. (1%)Pengfei Hong; Deepanway Ghosal; Navonil Majumder; Somak Aditya; Rada Mihalcea; Soujanya Poria Attack and Reset for Unlearning: Exploiting Adversarial Noise toward Machine Unlearning through Parameter Re-initialization. (1%)Yoonhwa Jung; Ikhyun Cho; Shun-Hsiang Hsu; Julia Hockenmaier 2024-01-16 Revealing Vulnerabilities in Stable Diffusion via Targeted Attacks. (99%)Chenyu Zhang; Lanjun Wang; Anan Liu Bag of Tricks to Boost Adversarial Transferability. (99%)Zeliang Zhang; Rongyi Zhu; Wei Yao; Xiaosen Wang; Chenliang Xu A Generative Adversarial Attack for Multilingual Text Classifiers. (99%)Tom Roth; Inigo Jauregi Unanue; Alsharif Abuadbba; Massimo Piccardi PPR: Enhancing Dodging Attacks while Maintaining Impersonation Attacks on Face Recognition Systems. (99%)Fengfan Zhou; Heifei Ling Robust Localization of Key Fob Using Channel Impulse Response of Ultra Wide Band Sensors for Keyless Entry Systems. (92%)Abhiram Kolli; Filippo Casamassima; Horst Possegger; Horst Bischof The Effect of Intrinsic Dataset Properties on Generalization: Unraveling Learning Differences Between Natural and Medical Images. (87%)Nicholas Konz; Maciej A. Mazurowski RandOhm: Mitigating Impedance Side-channel Attacks using Randomized Circuit Configurations. (9%)Saleh Khalaj Monfared; Domenic Forte; Shahin Tajik Towards Efficient and Certified Recovery from Poisoning Attacks in Federated Learning. (8%)Yu Jiang; Jiyuan Shen; Ziyao Liu; Chee Wei Tan; Kwok-Yan Lam IPR-NeRF: Ownership Verification meets Neural Radiance Field. (3%)Win Kent Ong; Kam Woh Ng; Chee Seng Chan; Yi Zhe Song; Tao Xiang IoTWarden: A Deep Reinforcement Learning Based Real-time Defense System to Mitigate Trigger-action IoT Attacks. (1%)Md Morshed Department of Software and Information Systems, University of North Carolina at Charlotte, Charlotte, USA Alam; Israt Department of Computer Science, University of Memphis, Memphis, USA Jahan; Weichao Department of Software and Information Systems, University of North Carolina at Charlotte, Charlotte, USA Wang 2024-01-15 Robustness Against Adversarial Attacks via Learning Confined Adversarial Polytopes. (99%)Shayan Mohajer Hamidi; Linfeng Ye Authorship Obfuscation in Multilingual Machine-Generated Text Detection. (12%)Dominik Macko; Robert Moro; Adaku Uchendu; Ivan Srba; Jason Samuel Lucas; Michiharu Yamashita; Nafis Irtiza Tripto; Dongwon Lee; Jakub Simko; Maria Bielikova 2024-01-14 LookAhead: Preventing DeFi Attacks via Unveiling Adversarial Contracts. (80%)Shoupeng Ren; Tianyu Tu; Jian Liu; Di Wu; Kui Ren Crafter: Facial Feature Crafting against Inversion-based Identity Theft on Deep Models. (70%)Shiming Wang; Zhe Ji; Liyao Xiang; Hao Zhang; Xinbing Wang; Chenghu Zhou; Bo Li 2024-01-13 Exploring Adversarial Attacks against Latent Diffusion Model from the Perspective of Adversarial Transferability. (99%)Junxi Chen; Junhao Dong; Xiaohua Xie Left-right Discrepancy for Adversarial Attack on Stereo Networks. (98%)Pengfei Wang; Xiaofei Hui; Beijia Lu; Nimrod Lilith; Jun Liu; Sameer Alam 2024-01-12 Adversarial Examples are Misaligned in Diffusion Model Manifolds. (98%)Peter Lorenz; Ricard Durall; Janis Keuper How Johnny Can Persuade LLMs to Jailbreak Them: Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs. (2%)Yi Zeng; Hongpeng Lin; Jingwen Zhang; Diyi Yang; Ruoxi Jia; Weiyan Shi Enhancing Consistency and Mitigating Bias: A Data Replay Approach for Incremental Learning. (1%)Chenyang Wang; Junjun Jiang; Xingyu Hu; Xianming Liu; Xiangyang Ji An Analytical Framework for Modeling and Synthesizing Malicious Attacks on ACC Vehicles. (1%)Shian Wang Intention Analysis Makes LLMs A Good Jailbreak Defender. (1%)Yuqi Zhang; Liang Ding; Lefei Zhang; Dacheng Tao 2024-01-11 GE-AdvGAN: Improving the transferability of adversarial samples by gradient editing-based adversarial generative model. (99%)Zhiyu Zhu; Huaming Chen; Xinyi Wang; Jiayu Zhang; Zhibo Jin; Kim-Kwang Raymond Choo; Jun Shen; Dong Yuan Universal Vulnerabilities in Large Language Models: In-context Learning Backdoor Attacks. (61%)Shuai Zhao; Meihuizi Jia; Luu Anh Tuan; Jinming Wen Open the Pandora's Box of LLMs: Jailbreaking LLMs through Representation Engineering. (22%)Tianlong Li; Shihan Dou; Wenhao Liu; Muling Wu; Changze Lv; Xiaoqing Zheng; Xuanjing Huang Can We Trust the Unlabeled Target Data? Towards Backdoor Attack and Defense on Model Adaptation. (8%)Lijun Sheng; Jian Liang; Ran He; Zilei Wang; Tieniu Tan Manipulating Feature Visualizations with Gradient Slingshots. (3%)Dilyara Bareeva; Marina M. -C. Höhne; Alexander Warnecke; Lukas Pirch; Klaus-Robert Müller; Konrad Rieck; Kirill Bykov Combating Adversarial Attacks with Multi-Agent Debate. (3%)Steffi Chern; Zhen Fan; Andy Liu 2024-01-10 Exploring Vulnerabilities of No-Reference Image Quality Assessment Models: A Query-Based Black-Box Method. (83%)Chenxi Yang; Yujia Liu; Dingquan Li; Tingting Jiang TrustLLM: Trustworthiness in Large Language Models. (75%)Lichao Sun; Yue Huang; Haoran Wang; Siyuan Wu; Qihui Zhang; Chujie Gao; Yixin Huang; Wenhan Lyu; Yixuan Zhang; Xiner Li; Zhengliang Liu; Yixin Liu; Yijue Wang; Zhikun Zhang; Bhavya Kailkhura; Caiming Xiong; Chaowei Xiao; Chunyuan Li; Eric Xing; Furong Huang; Hao Liu; Heng Ji; Hongyi Wang; Huan Zhang; Huaxiu Yao; Manolis Kellis; Marinka Zitnik; Meng Jiang; Mohit Bansal; James Zou; Jian Pei; Jian Liu; Jianfeng Gao; Jiawei Han; Jieyu Zhao; Jiliang Tang; Jindong Wang; John Mitchell; Kai Shu; Kaidi Xu; Kai-Wei Chang; Lifang He; Lifu Huang; Michael Backes; Neil Zhenqiang Gong; Philip S. Yu; Pin-Yu Chen; Quanquan Gu; Ran Xu; Rex Ying; Shuiwang Ji; Suman Jana; Tianlong Chen; Tianming Liu; Tianyi Zhou; Willian Wang; Xiang Li; Xiangliang Zhang; Xiao Wang; Xing Xie; Xun Chen; Xuyu Wang; Yan Liu; Yanfang Ye; Yinzhi Cao; Yong Chen; Yue Zhao SENet: Visual Detection of Online Social Engineering Attack Campaigns. (4%)Irfan Ozen; Karthika Subramani; Phani Vadrevu; Roberto Perdisci Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training. (2%)Evan Hubinger; Carson Denison; Jesse Mu; Mike Lambert; Meg Tong; Monte MacDiarmid; Tamera Lanham; Daniel M. Ziegler; Tim Maxwell; Newton Cheng; Adam Jermyn; Amanda Askell; Ansh Radhakrishnan; Cem Anil; David Duvenaud; Deep Ganguli; Fazl Barez; Jack Clark; Kamal Ndousse; Kshitij Sachan; Michael Sellitto; Mrinank Sharma; Nova DasSarma; Roger Grosse; Shauna Kravec; Yuntao Bai; Zachary Witten; Marina Favaro; Jan Brauner; Holden Karnofsky; Paul Christiano; Samuel R. Bowman; Logan Graham; Jared Kaplan; Sören Mindermann; Ryan Greenblatt; Buck Shlegeris; Nicholas Schiefer; Ethan Perez CoLafier: Collaborative Noisy Label Purifier With Local Intrinsic Dimensionality Guidance. (1%)Dongyu Zhang; Ruofan Hu; Elke Rundensteiner Brave: Byzantine-Resilient and Privacy-Preserving Peer-to-Peer Federated Learning. (1%)Zhangchen Xu; Fengqing Jiang; Luyao Niu; Jinyuan Jia; Radha Poovendran FBSDetector: Fake Base Station and Multi Step Attack Detection in Cellular Networks using Machine Learning. (1%)Kazi Samin Mubasshir; Imtiaz Karim; Elisa Bertino 2024-01-09 Revisiting Adversarial Training at Scale. (26%)Zeyu Wang; Xianhang Li; Hongru Zhu; Cihang Xie SoK: Facial Deepfake Detectors. (11%)Binh M. Le; Jiwon Kim; Shahroz Tariq; Kristen Moore; Alsharif Abuadbba; Simon S. Woo Advancing Ante-Hoc Explainable Models through Generative Adversarial Networks. (3%)Tanmay Garg; Deepika Vemuri; Vineeth N Balasubramanian 2024-01-08 Pre-trained Model Guided Fine-Tuning for Zero-Shot Adversarial Robustness. (99%)Sibo Wang; Jie Zhang; Zheng Yuan; Shiguang Shan Robustness Assessment of a Runway Object Classifier for Safe Aircraft Taxiing. (54%)Yizhak Elboher; Raya Elsaleh; Omri Isac; Mélanie Ducoffe; Audrey Galametz; Guillaume Povéda; Ryma Boumazouza; Noémie Cohen; Guy Katz Coupling Graph Neural Networks with Fractional Order Continuous Dynamics: A Robustness Study. (45%)Qiyu Kang; Kai Zhao; Yang Song; Yihang Xie; Yanan Zhao; Sijie Wang; Rui She; Wee Peng Tay Logits Poisoning Attack in Federated Distillation. (12%)Yuhan Tang; Zhiyuan Wu; Bo Gao; Tian Wen; Yuwei Wang; Sheng Sun Dense Hopfield Networks in the Teacher-Student Setting. (1%)Robin Thériault; Daniele Tantari 2024-01-07 Invisible Reflections: Leveraging Infrared Laser Reflections to Target Traffic Sign Perception. (87%)Takami Sato; Sri Hrushikesh Varma Bhupathiraju; Michael Clifford; Takeshi Sugawara; Qi Alfred Chen; Sara Rampazzi Data-Driven Subsampling in the Presence of an Adversarial Actor. (86%)Abu Shafin Mohammad Mahdee Jameel; Ahmed P. Mohamed; Jinho Yi; Aly El Gamal; Akshay Malhotra ROIC-DM: Robust Text Inference and Classification via Diffusion Model. (33%)Shilong Yuan; Wei Yuan; Hongzhi Yin; Tieke He 2024-01-06 Data-Dependent Stability Analysis of Adversarial Training. (98%)Yihan Wang; Shuang Liu; Xiao-Shan Gao End-to-End Anti-Backdoor Learning on Images and Time Series. (61%)Yujing Jiang; Xingjun Ma; Sarah Monazam Erfani; Yige Li; James Bailey 2024-01-05 Transferable Learned Image Compression-Resistant Adversarial Perturbations. (99%)Yang Sui; Zhuohang Li; Ding Ding; Xiang Pan; Xiaozhong Xu; Shan Liu; Zhenzhong Chen Enhancing targeted transferability via feature space fine-tuning. (98%)Hui Zeng; Biwei Chen; Anjie Peng Calibration Attack: A Framework For Adversarial Attacks Targeting Calibration. (68%)Stephen Obadinma; Xiaodan Zhu; Hongyu Guo A backdoor attack against link prediction tasks with graph neural networks. (38%)Jiazhu Dai; Haoyu Sun TEN-GUARD: Tensor Decomposition for Backdoor Attack Detection in Deep Neural Networks. (1%)Khondoker Murad Hossain; Tim Oates MLLM-Protector: Ensuring MLLM's Safety without Hurting Performance. (1%)Renjie Pi; Tianyang Han; Yueqi Xie; Rui Pan; Qing Lian; Hanze Dong; Jipeng Zhang; Tong Zhang 2024-01-04 Vulnerabilities Unveiled: Adversarially Attacking a Multimodal Vision Langauge Model for Pathology Imaging. (99%)Jai Prakash Veerla; Poojitha Thota; Partha Sai Guttikonda; Shirin Nilizadeh; Jacob M. Luber A Random Ensemble of Encrypted models for Enhancing Robustness against Adversarial Examples. (99%)Ryota Iijima; Sayaka Shiota; Hitoshi Kiya AdvSQLi: Generating Adversarial SQL Injections against Real-world WAF-as-a-service. (95%)Zhenqing Qu; Xiang Ling; Ting Wang; Xiang Chen; Shouling Ji; Chunming Wu Evasive Hardware Trojan through Adversarial Power Trace. (92%)Behnam Omidi; Khaled N. Khasawneh; Ihsen Alouani Object-oriented backdoor attack against image captioning. (76%)Meiling Li; Nan Zhong; Xinpeng Zhang; Zhenxing Qian; Sheng Li DEM: A Method for Certifying Deep Neural Network Classifier Outputs in Aerospace. (2%)Guy Katz; Natan Levy; Idan Refaeli; Raz Yerushalmi Secure Control of Connected and Automated Vehicles Using Trust-Aware Robust Event-Triggered Control Barrier Functions. (2%)H M Sabbir Ahmad; Ehsan Sabouni; Akua Dickson; Wei Xiao; Christos G. Cassandras; Wenchao Li 2024-01-03 Towards Robust Semantic Segmentation against Patch-based Attack via Attention Refinement. (92%)Zheng Yuan; Jie Zhang; Yude Wang; Shiguang Shan; Xilin Chen Spy-Watermark: Robust Invisible Watermarking for Backdoor Attack. (62%)Ruofei Wang; Renjie Wan; Zongyu Guo; Qing Guo; Rui Huang FullLoRA-AT: Efficiently Boosting the Robustness of Pretrained Vision Transformers. (33%)Zheng Yuan; Jie Zhang; Shiguang Shan Integrated Cyber-Physical Resiliency for Power Grids under IoT-Enabled Dynamic Botnet Attacks. (22%)Yuhan Zhao; Juntao Chen; Quanyan Zhu Enhancing Generalization of Invisible Facial Privacy Cloak via Gradient Accumulation. (1%)Xuannan Liu; Yaoyao Zhong; Weihong Deng; Hongzhi Shi; Xingchen Cui; Yunfeng Yin; Dongchao Wen 2024-01-02 JMA: a General Algorithm to Craft Nearly Optimal Targeted Adversarial Example. (99%)Benedetta Tondi; Wei Guo; Mauro Barni Dual Teacher Knowledge Distillation with Domain Alignment for Face Anti-spoofing. (92%)Zhe Kong; Wentian Zhang; Tao Wang; Kaihao Zhang; Yuexiang Li; Xiaoying Tang; Wenhan Luo SpecFormer: Guarding Vision Transformer Robustness via Maximum Singular Value Penalization. (64%)Xixu Hu; Runkai Zheng; Jindong Wang; Cheuk Hang Leung; Qi Wu; Xing Xie Unveiling the Stealthy Threat: Analyzing Slow Drift GPS Spoofing Attacks for Autonomous Vehicles in Urban Environments and Enabling the Resilience. (10%)Sagar Dasgupta; Abdullah Ahmed; Mizanur Rahman; Thejesh N. Bandi Imperio: Language-Guided Backdoor Attacks for Arbitrary Model Control. (4%)Ka-Ho Chow; Wenqi Wei; Lei Yu Will 6G be Semantic Communications? Opportunities and Challenges from Task Oriented and Secure Communications to Integrated Sensing. (2%)Yalin E. Sagduyu; Tugba Erpek; Aylin Yener; Sennur Ulukus 2024-01-01 Safety and Performance, Why Not Both? Bi-Objective Optimized Model Compression against Heterogeneous Attacks Toward AI Software Deployment. (12%)Jie Zhu; Leye Wang; Xiao Han; Anmin Liu; Tao Xie Detection and Defense Against Prominent Attacks on Preconditioned LLM-Integrated Virtual Assistants. (8%)Chun Fai Chan; Daniel Wankit Yip; Aysan Esmradi A Novel Evaluation Framework for Assessing Resilience Against Prompt Injection Attacks in Large Language Models. (2%)Daniel Wankit Yip; Aysan Esmradi; Chun Fai Chan 2023-12-31 AR-GAN: Generative Adversarial Network-Based Defense Method Against Adversarial Attacks on the Traffic Sign Classification System of Autonomous Vehicles. (99%)M Sabbir Salek; Abdullah Al Mamun; Mashrur Chowdhury Does Few-shot Learning Suffer from Backdoor Attacks? (98%)Xinwei Liu; Xiaojun Jia; Jindong Gu; Yuan Xun; Siyuan Liang; Xiaochun Cao Is It Possible to Backdoor Face Forgery Detection with Natural Triggers? (68%)Xiaoxuan Han; Songlin Yang; Wei Wang; Ziwen He; Jing Dong 2023-12-30 Explainability-Driven Leaf Disease Classification using Adversarial Training and Knowledge Distillation. (84%)Sebastian-Vasile Echim; Iulian-Marius Tăiatu; Dumitru-Clementin Cercel; Florin Pop CamPro: Camera-based Anti-Facial Recognition. (81%)Wenjun Zhu; Yuan Sun; Jiani Liu; Yushi Cheng; Xiaoyu Ji; Wenyuan Xu TPatch: A Triggered Physical Adversarial Patch. (76%)Wenjun Zhu; Xiaoyu Ji; Yushi Cheng; Shibo Zhang; Wenyuan Xu A clean-label graph backdoor attack method in node classification task. (9%)Xiaogang Xing; Ming Xu; Yujing Bai; Dongdong Yang 2023-12-29 Jatmo: Prompt Injection Defense by Task-Specific Finetuning. (54%)Julien Piet; Maha Alrashed; Chawin Sitawarin; Sizhe Chen; Zeming Wei; Elizabeth Sun; Basel Alomair; David Wagner Towards Faithful Explanations for Text Classification with Robustness Improvement and Explanation Guided Training. (9%)Dongfang Li; Baotian Hu; Qingcai Chen; Shan He 2023-12-28 Adversarial Attacks on Image Classification Models: Analysis and Defense. (99%)Jaydip Sen; Abhiraj Sen; Ananda Chatterjee BlackboxBench: A Comprehensive Benchmark of Black-box Adversarial Attacks. (99%)Meixi Zheng; Xuanchen Yan; Zihao Zhu; Hongrui Chen; Baoyuan Wu Attack Tree Analysis for Adversarial Evasion Attacks. (99%)Yuki Yamaguchi; Toshiaki Aoki Can you See me? On the Visibility of NOPs against Android Malware Detectors. (98%)Diego Soi; Davide Maiorca; Giorgio Giacinto; Harel Berger MVPatch: More Vivid Patch for Adversarial Camouflaged Attacks on Object Detectors in the Physical World. (98%)Zheng Zhou; Hongbo Zhao; Ju Liu; Qiaosheng Zhang; Liwei Geng; Shuchang Lyu; Wenquan Feng Explainability-Based Adversarial Attack on Graphs Through Edge Perturbation. (92%)Dibaloke Chanda; Saba Heidari Gheshlaghi; Nasim Yahya Soltani DOEPatch: Dynamically Optimized Ensemble Model for Adversarial Patches Generation. (83%)Wenyi Tan; Yang Li; Chenxing Zhao; Zhunga Liu; Quan Pan Securing NextG Systems against Poisoning Attacks on Federated Learning: A Game-Theoretic Solution. (64%)Yalin E. Sagduyu; Tugba Erpek; Yi Shi Timeliness: A New Design Metric and a New Attack Surface. (1%)Priyanka Kaswan; Sennur Ulukus 2023-12-27 Adversarial Attacks on LoRa Device Identification and Rogue Signal Detection with Deep Learning. (98%)Yalin E. Sagduyu; Tugba Erpek Domain Generalization with Vital Phase Augmentation. (3%)Ingyun Lee; Wooju Lee; Hyun Myung 2023-12-26 From Text to Multimodal: A Comprehensive Survey of Adversarial Example Generation in Question Answering Systems. (92%)Gulsum Yigit; Mehmet Fatih Amasyali Natural Adversarial Patch Generation Method Based on Latent Diffusion Model. (76%)Xianyi Chen; Fazhan Liu; Dong Jiang; Kai Yan Universal Pyramid Adversarial Training for Improved ViT Performance. (5%)Ping-yeh Chiang; Yipin Zhou; Omid Poursaeed; Satya Narayan Shukla; Ashish Shah; Tom Goldstein; Ser-Nam Lim Robust Survival Analysis with Adversarial Regularization. (4%)Michael Potter; Stefano Maxenti; Michael Everett 2023-12-25 GanFinger: GAN-Based Fingerprint Generation for Deep Neural Network Ownership Verification. (96%)Huali Ren; Anli Yan; Xiaojun Ren; Pei-Gen Ye; Chong-zhi Gao; Zhili Zhou; Jin Li Adversarial Item Promotion on Visually-Aware Recommender Systems by Guided Diffusion. (84%)Lijian Chen; Wei Yuan; Tong Chen; Guanhua Ye; Quoc Viet Hung Nguyen; Hongzhi Yin Punctuation Matters! Stealthy Backdoor Attack for Language Models. (11%)Xuan Sheng; Zhicheng Li; Zhaoyang Han; Xiangmao Chang; Piji Li 2023-12-23 Adversarial Data Poisoning for Fake News Detection: How to Make a Model Misclassify a Target News without Modifying It. (10%)Federico Siciliano; Luca Maiano; Lorenzo Papa; Federica Baccin; Irene Amerini; Fabrizio Silvestri Pre-trained Trojan Attacks for Visual Recognition. (1%)Aishan Liu; Xinwei Zhang; Yisong Xiao; Yuguang Zhou; Siyuan Liang; Jiakai Wang; Xianglong Liu; Xiaochun Cao; Dacheng Tao 2023-12-22 MEAOD: Model Extraction Attack against Object Detectors. (83%)Zeyu Li; Chenghui Shi; Yuwen Pu; Xuhong Zhang; Yu Li; Jinbao Li; Shouling Ji Asymmetric Bias in Text-to-Image Generation with Adversarial Attacks. (82%)Haz Sameen Shahgir; Xianghao Kong; Greg Ver Steeg; Yue Dong Understanding the Regularity of Self-Attention with Optimal Transport. (31%)Valérie Castin; Pierre Ablin; Gabriel Peyré Attacking Byzantine Robust Aggregation in High Dimensions. (22%)Sarthak Choudhary; Aashish Kolluri; Prateek Saxena SODA: Protecting Proprietary Information in On-Device Machine Learning Models. (4%)Akanksha Atrey; Ritwik Sinha; Saayan Mitra; Prashant Shenoy Energy-based learning algorithms for analog computing: a comparative study. (2%)Benjamin Scellier; Maxence Ernoult; Jack Kendall; Suhas Kumar Adaptive Domain Inference Attack. (1%)Yuechun Gu; Keke Chen 2023-12-21 AutoAugment Input Transformation for Highly Transferable Targeted Attacks. (99%)Haobo Lu; Xin Liu; Kun He Where and How to Attack? A Causality-Inspired Recipe for Generating Counterfactual Adversarial Examples. (98%)Ruichu Cai; Yuxuan Zhu; Jie Qiao; Zefeng Liang; Furui Liu; Zhifeng Hao Elevating Defenses: Bridging Adversarial Training and Watermarking for Model Resilience. (86%)Janvi Thakkar; Giulio Zizzo; Sergio Maffeis Adversarial Infrared Curves: An Attack on Infrared Pedestrian Detectors in the Physical World. (74%)Chengyin Hu; Weiwen Shi Exploiting Novel GPT-4 APIs. (8%)Kellin Pelrine; Mohammad Taufeeque; Michał Zając; Euan McLean; Adam Gleave 2023-12-20 Mutual-modality Adversarial Attack with Semantic Perturbation. (99%)Jingwen Ye; Ruonan Yu; Songhua Liu; Xinchao Wang LRS: Enhancing Adversarial Transferability through Lipschitz Regularized Surrogate. (99%)Tao Wu; Tie Luo; Donald C. Wunsch Adversarial Markov Games: On Adaptive Decision-Based Attacks and Defenses. (98%)Ilias Tsingenopoulos; Vera Rimmer; Davy Preuveneers; Fabio Pierazzi; Lorenzo Cavallaro; Wouter Joosen Benchmarking and Defending Against Indirect Prompt Injection Attacks on Large Language Models. (98%)Jingwei Yi; Yueqi Xie; Bin Zhu; Emre Kiciman; Guangzhong Sun; Xing Xie; Fangzhao Wu PGN: A perturbation generation network against deep reinforcement learning. (96%)Xiangjuan Li; Feifan Li; Yang Li; Quan Pan ARBiBench: Benchmarking Adversarial Robustness of Binarized Neural Networks. (96%)Peng Zhao; Jiehua Zhang; Bowen Peng; Longguang Wang; YingMei Wei; Yu Liu; Li Liu Scaling Compute Is Not All You Need for Adversarial Robustness. (93%)Edoardo Debenedetti; Zishen Wan; Maksym Andriushchenko; Vikash Sehwag; Kshitij Bhardwaj; Bhavya Kailkhura Doubly Perturbed Task Free Continual Learning. (9%)Byung Hyun Lee; Min-hwan Oh; Se Young Chun Interactive Visualization of Time-Varying Flow Fields Using Particle Tracing Neural Networks. (1%)Mengjiao Han; Jixian Li; Sudhanshu Sane; Shubham Gupta; Bei Wang; Steve Petruzza; Chris R. Johnson 2023-12-19 Tensor Train Decomposition for Adversarial Attacks on Computer Vision Models. (96%)Andrei Chertkov; Ivan Oseledets Trust, But Verify: A Survey of Randomized Smoothing Techniques. (78%)Anupriya Kumari; Devansh Bhardwaj; Sukrit Jindal; Sarthak Gupta SkyMask: Attack-agnostic Robust Federated Learning with Fine-grained Learnable Masks. (73%)Peishen Yan; Hao Wang; Tao Song; Yang Hua; Ruhui Ma; Ningxin Hu; Mohammad R. Haghighat; Haibing Guan Progressive Poisoned Data Isolation for Training-time Backdoor Defense. (61%)Yiming Chen; Haiwei Wu; Jiantao Zhou Adversarial AutoMixup. (11%)Huafeng Qin; Xin Jin; Yun Jiang; Mounim A. El-Yacoubi; Xinbo Gao I-CEE: Tailoring Explanations of Image Classifications Models to User Expertise. (1%)Yao Rong; Peizhu Qian; Vaibhav Unhelkar; Enkelejda Kasneci 2023-12-18 Gemini: A Family of Highly Capable Multimodal Models. (99%)Team Gemini; Rohan Anil; Sebastian Borgeaud; Yonghui Wu; Jean-Baptiste Alayrac; Jiahui Yu; Radu Soricut; Johan Schalkwyk; Andrew M. Dai; Anja Hauth; Katie Millican; David Silver; Slav Petrov; Melvin Johnson; Ioannis Antonoglou; Julian Schrittwieser; Amelia Glaese; Jilin Chen; Emily Pitler; Timothy Lillicrap; Angeliki Lazaridou; Orhan Firat; James Molloy; Michael Isard; Paul R. Barham; Tom Hennigan; Benjamin Lee; Fabio Viola; Malcolm Reynolds; Yuanzhong Xu; Ryan Doherty; Eli Collins; Clemens Meyer; Eliza Rutherford; Erica Moreira; Kareem Ayoub; Megha Goel; George Tucker; Enrique Piqueras; Maxim Krikun; Iain Barr; Nikolay Savinov; Ivo Danihelka; Becca Roelofs; Anaïs White; Anders Andreassen; Glehn Tamara von; Lakshman Yagati; Mehran Kazemi; Lucas Gonzalez; Misha Khalman; Jakub Sygnowski; Alexandre Frechette; Charlotte Smith; Laura Culp; Lev Proleev; Yi Luan; Xi Chen; James Lottes; Nathan Schucher; Federico Lebron; Alban Rrustemi; Natalie Clay; Phil Crone; Tomas Kocisky; Jeffrey Zhao; Bartek Perz; Dian Yu; Heidi Howard; Adam Bloniarz; Jack W. Rae; Han Lu; Laurent Sifre; Marcello Maggioni; Fred Alcober; Dan Garrette; Megan Barnes; Shantanu Thakoor; Jacob Austin; Gabriel Barth-Maron; William Wong; Rishabh Joshi; Rahma Chaabouni; Deeni Fatiha; Arun Ahuja; Ruibo Liu; Yunxuan Li; Sarah Cogan; Jeremy Chen; Chao Jia; Chenjie Gu; Qiao Zhang; Jordan Grimstad; Ale Jakse Hartman; Martin Chadwick; Gaurav Singh Tomar; Xavier Garcia; Evan Senter; Emanuel Taropa; Thanumalayan Sankaranarayana Pillai; Jacob Devlin; Michael Laskin; Diego de Las Casas; Dasha Valter; Connie Tao; Lorenzo Blanco; Adrià Puigdomènech Badia; David Reitter; Mianna Chen; Jenny Brennan; Clara Rivera; Sergey Brin; Shariq Iqbal; Gabriela Surita; Jane Labanowski; Abhi Rao; Stephanie Winkler; Emilio Parisotto; Yiming Gu; Kate Olszewska; Yujing Zhang; Ravi Addanki; Antoine Miech; Annie Louis; Laurent El Shafey; Denis Teplyashin; Geoff Brown; Elliot Catt; Nithya Attaluri; Jan Balaguer; Jackie Xiang; Pidong Wang; Zoe Ashwood; Anton Briukhov; Albert Webson; Sanjay Ganapathy; Smit Sanghavi; Ajay Kannan; Ming-Wei Chang; Axel Stjerngren; Josip Djolonga; Yuting Sun; Ankur Bapna; Matthew Aitchison; Pedram Pejman; Henryk Michalewski; Tianhe Yu; Cindy Wang; Juliette Love; Junwhan Ahn; Dawn Bloxwich; Kehang Han; Peter Humphreys; Thibault Sellam; James Bradbury; Varun Godbole; Sina Samangooei; Bogdan Damoc; Alex Kaskasoli; Sébastien M. R. Arnold; Vijay Vasudevan; Shubham Agrawal; Jason Riesa; Dmitry Lepikhin; Richard Tanburn; Srivatsan Srinivasan; Hyeontaek Lim; Sarah Hodkinson; Pranav Shyam; Johan Ferret; Steven Hand; Ankush Garg; Tom Le Paine; Jian Li; Yujia Li; Minh Giang; Alexander Neitz; Zaheer Abbas; Sarah York; Machel Reid; Elizabeth Cole; Aakanksha Chowdhery; Dipanjan Das; Dominika Rogozińska; Vitaly Nikolaev; Pablo Sprechmann; Zachary Nado; Lukas Zilka; Flavien Prost; Luheng He; Marianne Monteiro; Gaurav Mishra; Chris Welty; Josh Newlan; Dawei Jia; Miltiadis Allamanis; Clara Huiyi Hu; Liedekerke Raoul de; Justin Gilmer; Carl Saroufim; Shruti Rijhwani; Shaobo Hou; Disha Shrivastava; Anirudh Baddepudi; Alex Goldin; Adnan Ozturel; Albin Cassirer; Yunhan Xu; Daniel Sohn; Devendra Sachan; Reinald Kim Amplayo; Craig Swanson; Dessie Petrova; Shashi Narayan; Arthur Guez; Siddhartha Brahma; Jessica Landon; Miteyan Patel; Ruizhe Zhao; Kevin Villela; Luyu Wang; Wenhao Jia; Matthew Rahtz; Mai Giménez; Legg Yeung; Hanzhao Lin; James Keeling; Petko Georgiev; Diana Mincu; Boxi Wu; Salem Haykal; Rachel Saputro; Kiran Vodrahalli; James Qin; Zeynep Cankara; Abhanshu Sharma; Nick Fernando; Will Hawkins; Behnam Neyshabur; Solomon Kim; Adrian Hutter; Priyanka Agrawal; Alex Castro-Ros; George van den Driessche; Tao Wang; Fan Yang; Shuo-yiin Chang; Paul Komarek; Ross McIlroy; Mario Lučić; Guodong Zhang; Wael Farhan; Michael Sharman; Paul Natsev; Paul Michel; Yong Cheng; Yamini Bansal; Siyuan Qiao; Kris Cao; Siamak Shakeri; Christina Butterfield; Justin Chung; Paul Kishan Rubenstein; Shivani Agrawal; Arthur Mensch; Kedar Soparkar; Karel Lenc; Timothy Chung; Aedan Pope; Loren Maggiore; Jackie Kay; Priya Jhakra; Shibo Wang; Joshua Maynez; Mary Phuong; Taylor Tobin; Andrea Tacchetti; Maja Trebacz; Kevin Robinson; Yash Katariya; Sebastian Riedel; Paige Bailey; Kefan Xiao; Nimesh Ghelani; Lora Aroyo; Ambrose Slone; Neil Houlsby; Xuehan Xiong; Zhen Yang; Elena Gribovskaya; Jonas Adler; Mateo Wirth; Lisa Lee; Music Li; Thais Kagohara; Jay Pavagadhi; Sophie Bridgers; Anna Bortsova; Sanjay Ghemawat; Zafarali Ahmed; Tianqi Liu; Richard Powell; Vijay Bolina; Mariko Iinuma; Polina Zablotskaia; James Besley; Da-Woon Chung; Timothy Dozat; Ramona Comanescu; Xiance Si; Jeremy Greer; Guolong Su; Martin Polacek; Raphaël Lopez Kaufman; Simon Tokumine; Hexiang Hu; Elena Buchatskaya; Yingjie Miao; Mohamed Elhawaty; Aditya Siddhant; Nenad Tomasev; Jinwei Xing; Christina Greer; Helen Miller; Shereen Ashraf; Aurko Roy; Zizhao Zhang; Ada Ma; Angelos Filos; Milos Besta; Rory Blevins; Ted Klimenko; Chih-Kuan Yeh; Soravit Changpinyo; Jiaqi Mu; Oscar Chang; Mantas Pajarskas; Carrie Muir; Vered Cohen; Charline Le Lan; Krishna Haridasan; Amit Marathe; Steven Hansen; Sholto Douglas; Rajkumar Samuel; Mingqiu Wang; Sophia Austin; Chang Lan; Jiepu Jiang; Justin Chiu; Jaime Alonso Lorenzo; Lars Lowe Sjösund; Sébastien Cevey; Zach Gleicher; Thi Avrahami; Anudhyan Boral; Hansa Srinivasan; Vittorio Selo; Rhys May; Konstantinos Aisopos; Léonard Hussenot; Livio Baldini Soares; Kate Baumli; Michael B. Chang; Adrià Recasens; Ben Caine; Alexander Pritzel; Filip Pavetic; Fabio Pardo; Anita Gergely; Justin Frye; Vinay Ramasesh; Dan Horgan; Kartikeya Badola; Nora Kassner; Subhrajit Roy; Ethan Dyer; Víctor Campos; Alex Tomala; Yunhao Tang; Dalia El Badawy; Elspeth White; Basil Mustafa; Oran Lang; Abhishek Jindal; Sharad Vikram; Zhitao Gong; Sergi Caelles; Ross Hemsley; Gregory Thornton; Fangxiaoyu Feng; Wojciech Stokowiec; Ce Zheng; Phoebe Thacker; Çağlar Ünlü; Zhishuai Zhang; Mohammad Saleh; James Svensson; Max Bileschi; Piyush Patil; Ankesh Anand; Roman Ring; Katerina Tsihlas; Arpi Vezer; Marco Selvi; Toby Shevlane; Mikel Rodriguez; Tom Kwiatkowski; Samira Daruki; Keran Rong; Allan Dafoe; Nicholas FitzGerald; Keren Gu-Lemberg; Mina Khan; Lisa Anne Hendricks; Marie Pellat; Vladimir Feinberg; James Cobon-Kerr; Tara Sainath; Maribeth Rauh; Sayed Hadi Hashemi; Richard Ives; Yana Hasson; YaGuang Li; Eric Noland; Yuan Cao; Nathan Byrd; Le Hou; Qingze Wang; Thibault Sottiaux; Michela Paganini; Jean-Baptiste Lespiau; Alexandre Moufarek; Samer Hassan; Kaushik Shivakumar; Amersfoort Joost van; Amol Mandhane; Pratik Joshi; Anirudh Goyal; Matthew Tung; Andrew Brock; Hannah Sheahan; Vedant Misra; Cheng Li; Nemanja Rakićević; Mostafa Dehghani; Fangyu Liu; Sid Mittal; Junhyuk Oh; Seb Noury; Eren Sezener; Fantine Huot; Matthew Lamm; Cao Nicola De; Charlie Chen; Gamaleldin Elsayed; Ed Chi; Mahdis Mahdieh; Ian Tenney; Nan Hua; Ivan Petrychenko; Patrick Kane; Dylan Scandinaro; Rishub Jain; Jonathan Uesato; Romina Datta; Adam Sadovsky; Oskar Bunyan; Dominik Rabiej; Shimu Wu; John Zhang; Gautam Vasudevan; Edouard Leurent; Mahmoud Alnahlawi; Ionut Georgescu; Nan Wei; Ivy Zheng; Betty Chan; Pam G Rabinovitch; Piotr Stanczyk; Ye Zhang; David Steiner; Subhajit Naskar; Michael Azzam; Matthew Johnson; Adam Paszke; Chung-Cheng Chiu; Jaume Sanchez Elias; Afroz Mohiuddin; Faizan Muhammad; Jin Miao; Andrew Lee; Nino Vieillard; Sahitya Potluri; Jane Park; Elnaz Davoodi; Jiageng Zhang; Jeff Stanway; Drew Garmon; Abhijit Karmarkar; Zhe Dong; Jong Lee; Aviral Kumar; Luowei Zhou; Jonathan Evens; William Isaac; Zhe Chen; Johnson Jia; Anselm Levskaya; Zhenkai Zhu; Chris Gorgolewski; Peter Grabowski; Yu Mao; Alberto Magni; Kaisheng Yao; Javier Snaider; Norman Casagrande; Paul Suganthan; Evan Palmer; Geoffrey Irving; Edward Loper; Manaal Faruqui; Isha Arkatkar; Nanxin Chen; Izhak Shafran; Michael Fink; Alfonso Castaño; Irene Giannoumis; Wooyeol Kim; Mikołaj Rybiński; Ashwin Sreevatsa; Jennifer Prendki; David Soergel; Adrian Goedeckemeyer; Willi Gierke; Mohsen Jafari; Meenu Gaba; Jeremy Wiesner; Diana Gage Wright; Yawen Wei; Harsha Vashisht; Yana Kulizhskaya; Jay Hoover; Maigo Le; Lu Li; Chimezie Iwuanyanwu; Lu Liu; Kevin Ramirez; Andrey Khorlin; Albert Cui; Tian LIN; Marin Georgiev; Marcus Wu; Ricardo Aguilar; Keith Pallo; Abhishek Chakladar; Alena Repina; Xihui Wu; der Weide Tom van; Priya Ponnapalli; Caroline Kaplan; Jiri Simsa; Shuangfeng Li; Olivier Dousse; Fan Yang; Jeff Piper; Nathan Ie; Minnie Lui; Rama Pasumarthi; Nathan Lintz; Anitha Vijayakumar; Lam Nguyen Thiet; Daniel Andor; Pedro Valenzuela; Cosmin Paduraru; Daiyi Peng; Katherine Lee; Shuyuan Zhang; Somer Greene; Duc Dung Nguyen; Paula Kurylowicz; Sarmishta Velury; Sebastian Krause; Cassidy Hardin; Lucas Dixon; Lili Janzer; Kiam Choo; Ziqiang Feng; Biao Zhang; Achintya Singhal; Tejasi Latkar; Mingyang Zhang; Quoc Le; Elena Allica Abellan; Dayou Du; Dan McKinnon; Natasha Antropova; Tolga Bolukbasi; Orgad Keller; David Reid; Daniel Finchelstein; Maria Abi Raad; Remi Crocker; Peter Hawkins; Robert Dadashi; Colin Gaffney; Sid Lall; Ken Franko; Egor Filonov; Anna Bulanova; Rémi Leblond; Vikas Yadav; Shirley Chung; Harry Askham; Luis C. Cobo; Kelvin Xu; Felix Fischer; Jun Xu; Christina Sorokin; Chris Alberti; Chu-Cheng Lin; Colin Evans; Hao Zhou; Alek Dimitriev; Hannah Forbes; Dylan Banarse; Zora Tung; Jeremiah Liu; Mark Omernick; Colton Bishop; Chintu Kumar; Rachel Sterneck; Ryan Foley; Rohan Jain; Swaroop Mishra; Jiawei Xia; Taylor Bos; Geoffrey Cideron; Ehsan Amid; Francesco Piccinno; Xingyu Wang; Praseem Banzal; Petru Gurita; Hila Noga; Premal Shah; Daniel J. Mankowitz; Alex Polozov; Nate Kushman; Victoria Krakovna; Sasha Brown; MohammadHossein Bateni; Dennis Duan; Vlad Firoiu; Meghana Thotakuri; Tom Natan; Anhad Mohananey; Matthieu Geist; Sidharth Mudgal; Sertan Girgin; Hui Li; Jiayu Ye; Ofir Roval; Reiko Tojo; Michael Kwong; James Lee-Thorp; Christopher Yew; Quan Yuan; Sumit Bagri; Danila Sinopalnikov; Sabela Ramos; John Mellor; Abhishek Sharma; Aliaksei Severyn; Jonathan Lai; Kathy Wu; Heng-Tze Cheng; David Miller; Nicolas Sonnerat; Denis Vnukov; Rory Greig; Jennifer Beattie; Emily Caveness; Libin Bai; Julian Eisenschlos; Alex Korchemniy; Tomy Tsai; Mimi Jasarevic; Weize Kong; Phuong Dao; Zeyu Zheng; Frederick Liu; Fan Yang; Rui Zhu; Mark Geller; Tian Huey Teh; Jason Sanmiya; Evgeny Gladchenko; Nejc Trdin; Andrei Sozanschi; Daniel Toyama; Evan Rosen; Sasan Tavakkol; Linting Xue; Chen Elkind; Oliver Woodman; John Carpenter; George Papamakarios; Rupert Kemp; Sushant Kafle; Tanya Grunina; Rishika Sinha; Alice Talbert; Abhimanyu Goyal; Diane Wu; Denese Owusu-Afriyie; Cosmo Du; Chloe Thornton; Jordi Pont-Tuset; Pradyumna Narayana; Jing Li; Sabaer Fatehi; John Wieting; Omar Ajmeri; Benigno Uria; Tao Zhu; Yeongil Ko; Laura Knight; Amélie Héliou; Ning Niu; Shane Gu; Chenxi Pang; Dustin Tran; Yeqing Li; Nir Levine; Ariel Stolovich; Norbert Kalb; Rebeca Santamaria-Fernandez; Sonam Goenka; Wenny Yustalim; Robin Strudel; Ali Elqursh; Balaji Lakshminarayanan; Charlie Deck; Shyam Upadhyay; Hyo Lee; Mike Dusenberry; Zonglin Li; Xuezhi Wang; Kyle Levin; Raphael Hoffmann; Dan Holtmann-Rice; Olivier Bachem; Summer Yue; Sho Arora; Eric Malmi; Daniil Mirylenka; Qijun Tan; Christy Koh; Soheil Hassas Yeganeh; Siim Põder; Steven Zheng; Francesco Pongetti; Mukarram Tariq; Yanhua Sun; Lucian Ionita; Mojtaba Seyedhosseini; Pouya Tafti; Ragha Kotikalapudi; Zhiyu Liu; Anmol Gulati; Jasmine Liu; Xinyu Ye; Bart Chrzaszcz; Lily Wang; Nikhil Sethi; Tianrun Li; Ben Brown; Shreya Singh; Wei Fan; Aaron Parisi; Joe Stanton; Chenkai Kuang; Vinod Koverkathu; Christopher A. Choquette-Choo; Yunjie Li; TJ Lu; Abe Ittycheriah; Prakash Shroff; Pei Sun; Mani Varadarajan; Sanaz Bahargam; Rob Willoughby; David Gaddy; Ishita Dasgupta; Guillaume Desjardins; Marco Cornero; Brona Robenek; Bhavishya Mittal; Ben Albrecht; Ashish Shenoy; Fedor Moiseev; Henrik Jacobsson; Alireza Ghaffarkhah; Morgane Rivière; Alanna Walton; Clément Crepy; Alicia Parrish; Yuan Liu; Zongwei Zhou; Clement Farabet; Carey Radebaugh; Praveen Srinivasan; der Salm Claudia van; Andreas Fidjeland; Salvatore Scellato; Eri Latorre-Chimoto; Hanna Klimczak-Plucińska; David Bridson; Cesare Dario de; Tom Hudson; Piermaria Mendolicchio; Lexi Walker; Alex Morris; Ivo Penchev; Matthew Mauger; Alexey Guseynov; Alison Reid; Seth Odoom; Lucia Loher; Victor Cotruta; Madhavi Yenugula; Dominik Grewe; Anastasia Petrushkina; Tom Duerig; Antonio Sanchez; Steve Yadlowsky; Amy Shen; Amir Globerson; Adam Kurzrok; Lynette Webb; Sahil Dua; Dong Li; Preethi Lahoti; Surya Bhupatiraju; Dan Hurt; Haroon Qureshi; Ananth Agarwal; Tomer Shani; Matan Eyal; Anuj Khare; Shreyas Rammohan Belle; Lei Wang; Chetan Tekur; Mihir Sanjay Kale; Jinliang Wei; Ruoxin Sang; Brennan Saeta; Tyler Liechty; Yi Sun; Yao Zhao; Stephan Lee; Pandu Nayak; Doug Fritz; Manish Reddy Vuyyuru; John Aslanides; Nidhi Vyas; Martin Wicke; Xiao Ma; Taylan Bilal; Evgenii Eltyshev; Daniel Balle; Nina Martin; Hardie Cate; James Manyika; Keyvan Amiri; Yelin Kim; Xi Xiong; Kai Kang; Florian Luisier; Nilesh Tripuraneni; David Madras; Mandy Guo; Austin Waters; Oliver Wang; Joshua Ainslie; Jason Baldridge; Han Zhang; Garima Pruthi; Jakob Bauer; Feng Yang; Riham Mansour; Jason Gelman; Yang Xu; George Polovets; Ji Liu; Honglong Cai; Warren Chen; XiangHai Sheng; Emily Xue; Sherjil Ozair; Adams Yu; Christof Angermueller; Xiaowei Li; Weiren Wang; Julia Wiesinger; Emmanouil Koukoumidis; Yuan Tian; Anand Iyer; Madhu Gurumurthy; Mark Goldenson; Parashar Shah; MK Blake; Hongkun Yu; Anthony Urbanowicz; Jennimaria Palomaki; Chrisantha Fernando; Kevin Brooks; Ken Durden; Harsh Mehta; Nikola Momchev; Elahe Rahimtoroghi; Maria Georgaki; Amit Raul; Sebastian Ruder; Morgan Redshaw; Jinhyuk Lee; Komal Jalan; Dinghua Li; Ginger Perng; Blake Hechtman; Parker Schuh; Milad Nasr; Mia Chen; Kieran Milan; Vladimir Mikulik; Trevor Strohman; Juliana Franco; Tim Green; Demis Hassabis; Koray Kavukcuoglu; Jeffrey Dean; Oriol Vinyals Adv-Diffusion: Imperceptible Adversarial Face Identity Attack via Latent Diffusion Model. (99%)Decheng Liu; Xijun Wang; Chunlei Peng; Nannan Wang; Ruiming Hu; Xinbo Gao The Ultimate Combo: Boosting Adversarial Example Transferability by Composing Data Augmentations. (99%)Zebin Yun; Achi-Or Weingarten; Eyal Ronen; Mahmood Sharif DataElixir: Purifying Poisoned Dataset to Mitigate Backdoor Attacks via Diffusion Models. (16%)Jiachen Zhou; Peizhuo Lv; Yibing Lan; Guozhu Meng; Kai Chen; Hualong Ma A Comprehensive Survey of Attack Techniques, Implementation, and Mitigation Strategies in Large Language Models. (10%)Aysan Esmradi; Daniel Wankit Yip; Chun Fai Chan Model Stealing Attack against Recommender System. (10%)Zhihao Zhu; Rui Fan; Chenwang Wu; Yi Yang; Defu Lian; Enhong Chen Model Stealing Attack against Graph Classification with Authenticity, Uncertainty and Diversity. (4%)Zhihao Zhu; Chenwang Wu; Rui Fan; Yi Yang; Defu Lian; Enhong Chen MISA: Unveiling the Vulnerabilities in Split Federated Learning. (1%)Wei Wan; Yuxuan Ning; Shengshan Hu; Lulu Xue; Minghui Li; Leo Yu Zhang; Hai Jin A Survey of Side-Channel Attacks in Context of Cache -- Taxonomies, Analysis and Mitigation. (1%)Ankit Pulkit; Smita Naval; Vijay Laxmi 2023-12-17 UltraClean: A Simple Framework to Train Robust Neural Networks against Backdoor Attacks. (98%)Bingyin Zhao; Yingjie Lao The Pros and Cons of Adversarial Robustness. (92%)Yacine Izza; Joao Marques-Silva A Mutation-Based Method for Multi-Modal Jailbreaking Attack Detection. (80%)Xiaoyu Zhang; Cen Zhang; Tianlin Li; Yihao Huang; Xiaojun Jia; Xiaofei Xie; Yang Liu; Chao Shen Robust Node Representation Learning via Graph Variational Diffusion Networks. (11%)Jun Zhuang; Mohammad Al Hasan A Study on Transferability of Deep Learning Models for Network Intrusion Detection. (4%)Shreya Ghosh; Abu Shafin Mohammad Mahdee Jameel; Aly El Gamal 2023-12-16 Perturbation-Invariant Adversarial Training for Neural Ranking Models: Improving the Effectiveness-Robustness Trade-Off. (99%)Yu-An Liu; Ruqing Zhang; Mingkun Zhang; Wei Chen; Rijke Maarten de; Jiafeng Guo; Xueqi Cheng Rethinking Robustness of Model Attributions. (80%)Sandesh Kamath; Sankalp Mittal; Amit Deshpande; Vineeth N Balasubramanian SAME: Sample Reconstruction Against Model Extraction Attacks. (13%)Yi Xie; Jie Zhang; Shiqian Zhao; Tianwei Zhang; Xiaofeng Chen TrojFair: Trojan Fairness Attacks. (8%)Mengxin Zheng; Jiaqi Xue; Yi Sheng; Lei Yang; Qian Lou; Lei Jiang Transformers in Unsupervised Structure-from-Motion. (3%)Hemang Chawla; Arnav Varma; Elahe Arani; Bahram Zonooz TrojFSP: Trojan Insertion in Few-shot Prompt Tuning. (2%)Mengxin Zheng; Jiaqi Xue; Xun Chen; YanShan Wang; Qian Lou; Lei Jiang 2023-12-15 LogoStyleFool: Vitiating Video Recognition Systems via Logo Style Transfer. (99%)Yuxin Cao; Ziyu Zhao; Xi Xiao; Derui Wang; Minhui Xue; Jin Lu Embodied Adversarial Attack: A Dynamic Robust Physical Attack in Autonomous Driving. (99%)Yitong Sun; Yao Huang; Xingxing Wei Towards Transferable Targeted 3D Adversarial Attack in the Physical World. (99%)Yao Huang; Yinpeng Dong; Shouwei Ruan; Xiao Yang; Hang Su; Xingxing Wei A Malware Classification Survey on Adversarial Attacks and Defences. (98%)Mahesh Datta Sai Ponnuru; Likhitha Amasala; Tanu Sree Bhimavarapu; Guna Chaitanya Garikipati FlowMur: A Stealthy and Practical Audio Backdoor Attack with Limited Knowledge. (76%)Jiahe Lan; Jie Wang; Baochen Yan; Zheng Yan; Elisa Bertino Closing the Gap: Achieving Better Accuracy-Robustness Tradeoffs Against Query-Based Attacks. (74%)Pascal Zimmer; Sébastien Andreina; Giorgia Azzurra Marson; Ghassan Karame Fragility, Robustness and Antifragility in Deep Learning. (67%)Chandresh Pravin; Ivan Martino; Giuseppe Nicosia; Varun Ojha Verification-Friendly Deep Neural Networks. (56%)Anahita Baninajjar; Ahmed Rezine; Amir Aminifar Silent Guardian: Protecting Text from Malicious Exploitation by Large Language Models. (8%)Jiawei Zhao; Kejiang Chen; Xiaojian Yuan; Yuang Qi; Weiming Zhang; Nenghai Yu 2023-12-14 AVA: Inconspicuous Attribute Variation-based Adversarial Attack bypassing DeepFake Detection. (99%)Xiangtao Meng; Li Wang; Shanqing Guo; Lei Ju; Qingchuan Zhao Continual Adversarial Defense. (95%)Qian Wang; Yaoyao Liu; Hefei Ling; Yingwei Li; Qihao Liu; Ping Li; Jiazhong Chen; Alan Yuille; Ning Yu SlowTrack: Increasing the Latency of Camera-based Perception in Autonomous Driving Using Adversarial Examples. (92%)Chen Ma; Ningfei Wang; Qi Alfred Chen; Chao Shen On the Difficulty of Defending Contrastive Learning against Backdoor Attacks. (84%)Changjiang Li; Ren Pang; Bochuan Cao; Zhaohan Xi; Jinghui Chen; Shouling Ji; Ting Wang Detection and Defense of Unlearnable Examples. (81%)Yifan Zhu; Lijia Yu; Xiao-Shan Gao Improve Robustness of Reinforcement Learning against Observation Perturbations via $l_\infty$ Lipschitz Policy Networks. (81%)Buqing Nie; Jingtian Ji; Yangqing Fu; Yue Gao Adversarial Robustness on Image Classification with $k$-means. (81%)Rollin Omari; Junae Kim; Paul Montague Data and Model Poisoning Backdoor Attacks on Wireless Federated Learning, and the Defense Mechanisms: A Comprehensive Survey. (76%)Yichen Wan; Youyang Qu; Wei Ni; Yong Xiang; Longxiang Gao; Ekram Hossain DRAM-Locker: A General-Purpose DRAM Protection Mechanism against Adversarial DNN Weight Attacks. (45%)Ranyang Zhou; Sabbir Ahmed; Arman Roohi; Adnan Siraj Rakin; Shaahin Angizi No-Skim: Towards Efficiency Robustness Evaluation on Skimming-based Language Models. (45%)Shengyao Zhang; Mi Zhang; Xudong Pan; Min Yang Forbidden Facts: An Investigation of Competing Objectives in Llama-2. (45%)Tony T. Wang; Miles Wang; Kaivalya Hariharan; Nir Shavit Coevolutionary Algorithm for Building Robust Decision Trees under Minimax Regret. (13%)Adam Żychowski; Andrew Perrault; Jacek Mańdziuk Exploring Transferability for Randomized Smoothing. (5%)Kai Qiu; Huishuai Zhang; Zhirong Wu; Stephen Lin Split-Ensemble: Efficient OOD-aware Ensemble via Task and Model Splitting. (1%)Anthony Chen; Huanrui Yang; Yulu Gan; Denis A Gudovskiy; Zhen Dong; Haofan Wang; Tomoyuki Okuno; Yohei Nakata; Shanghang Zhang; Kurt Keutzer 2023-12-13 Defenses in Adversarial Machine Learning: A Survey. (99%)Baoyuan Wu; Shaokui Wei; Mingli Zhu; Meixi Zheng; Zihao Zhu; Mingda Zhang; Hongrui Chen; Danni Yuan; Li Liu; Qingshan Liu Robust Few-Shot Named Entity Recognition with Boundary Discrimination and Correlation Purification. (99%)Xiaojun Xue; Chunxia Zhang; Tianxiang Xu; Zhendong Niu Universal Adversarial Framework to Improve Adversarial Robustness for Diabetic Retinopathy Detection. (98%)Samrat Mukherjee; Dibyanayan Bandyopadhyay; Baban Gain; Asif Ekbal Towards Inductive Robustness: Distilling and Fostering Wave-induced Resonance in Transductive GCNs Against Graph Adversarial Attacks. (83%)Ao Liu; Wenshan Li; Tao Li; Beibei Li; Hanyuan Huang; Pan Zhou Scalable Ensemble-based Detection Method against Adversarial Attacks for speaker verification. (64%)Haibin Wu; Heng-Cheng Kuo; Yu Tsao; Hung-yi Lee Accelerating the Global Aggregation of Local Explanations. (47%)Alon Mor; Yonatan Belinkov; Benny Kimelfeld Erasing Self-Supervised Learning Backdoor by Cluster Activation Masking. (13%)Shengsheng Qian; Yifei Wang; Dizhan Xue; Shengjie Zhang; Huaiwen Zhang; Changsheng Xu Efficient Representation of the Activation Space in Deep Neural Networks. (11%)Tanya Akumu; Celia Cintas; Girmaw Abebe Tadesse; Adebayo Oshingbesan; Skyler Speakman; Edward III McFowland Efficient Toxic Content Detection by Bootstrapping and Distilling Large Language Models. (1%)Jiang Zhang; Qiong Wu; Yiming Xu; Cheng Cao; Zheng Du; Konstantinos Psounis 2023-12-12 Radio Signal Classification by Adversarially Robust Quantum Machine Learning. (99%)Yanqiu Wu; Eromanga Adermann; Chandra Thapa; Seyit Camtepe; Hajime Suzuki; Muhammad Usman SSTA: Salient Spatially Transformed Attack. (99%)Renyang Liu; Wei Zhou; Sixin Wu; Jun Zhao; Kwok-Yan Lam DTA: Distribution Transform-based Attack for Query-Limited Scenario. (99%)Renyang Liu; Wei Zhou; Xin Jin; Song Gao; Yuanyu Wang; Ruxin Wang May the Noise be with you: Adversarial Training without Adversarial Examples. (98%)Ayoub Arous; Andres F Lopez-Lopera; Nael Abu-Ghazaleh; Ihsen Alouani Collapse-Oriented Adversarial Training with Triplet Decoupling for Robust Image Retrieval. (98%)Qiwei Tian; Chenhao Lin; Qian Li; Zhengyu Zhao; Chao Shen Focus on Hiders: Exploring Hidden Threats for Enhancing Adversarial Training. (98%)Qian Li; Yuxiao Hu; Yinpeng Dong; Dongxiao Zhang; Yuntian Chen QuadAttack: A Quadratic Programming Approach to Ordered Top-K Attacks. (97%)Thomas Paniagua; Ryan Grainger; Tianfu Wu Attacking the Loop: Adversarial Attacks on Graph-based Loop Closure Detection. (92%)Jonathan J. Y. Kim; Martin Urschler; Patricia J. Riddle; Jorg S. Wicker ReRoGCRL: Representation-based Robustness in Goal-Conditioned Reinforcement Learning. (86%)Xiangyu Yin; Sihao Wu; Jiaxu Liu; Meng Fang; Xingyu Zhao; Xiaowei Huang; Wenjie Ruan Robust MRI Reconstruction by Smoothed Unrolling (SMUG). (82%)Shijun Liang; Van Hoang Minh Nguyen; Jinghan Jia; Ismail Alkhouri; Sijia Liu; Saiprasad Ravishankar Cost Aware Untargeted Poisoning Attack against Graph Neural Networks,. (70%)Yuwei Han; Yuni Lai; Yulin Zhu; Kai Zhou EdgePruner: Poisoned Edge Pruning in Graph Contrastive Learning. (47%)Hiroya Kato; Kento Hasegawa; Seira Hidano; Kazuhide Fukushima Causality Analysis for Evaluating the Security of Large Language Models. (22%)Wei Zhao; Zhe Li; Jun Sun SimAC: A Simple Anti-Customization Method against Text-to-Image Synthesis of Diffusion Models. (13%)Feifei Wang; Zhentao Tan; Tianyi Wei; Yue Wu; Qidong Huang Divide-and-Conquer Attack: Harnessing the Power of LLM to Bypass Safety Filters of Text-to-Image Models. (8%)Yimo Deng; Huangxun Chen Eroding Trust In Aerial Imagery: Comprehensive Analysis and Evaluation Of Adversarial Attacks In Geospatial Systems. (5%)Michael Lanier; Aayush Dhakal; Zhexiao Xiong; Arthur Li; Nathan Jacobs; Yevgeniy Vorobeychik Securing Graph Neural Networks in MLaaS: A Comprehensive Realization of Query-based Integrity Verification. (2%)Bang Wu; Xingliang Yuan; Shuo Wang; Qi Li; Minhui Xue; Shirui Pan Majority is Not Required: A Rational Analysis of the Private Double-Spend Attack from a Sub-Majority Adversary. (1%)Yanni Georghiades; Rajesh Mishra; Karl Kreder; Sriram Vishwanath 2023-12-11 Towards Transferable Adversarial Attacks with Centralized Perturbation. (99%)Shangbo Wu; Yu-an Tan; Yajie Wang; Ruinan Ma; Wencong Ma; Yuanzhang Li MalPurifier: Enhancing Android Malware Detection with Adversarial Purification against Evasion Attacks. (98%)Yuyang Zhou; Guang Cheng; Zongyao Chen; Shui Yu Sparse but Strong: Crafting Adversarially Robust Graph Lottery Tickets. (83%)Subhajit Dutta Chowdhury; Zhiyu Ni; Qingyuan Peng; Souvik Kundu; Pierluigi Nuzzo Reward Certification for Policy Smoothed Reinforcement Learning. (78%)Ronghui Mu; Leandro Soriano Marcolino; Tianle Zhang; Yanghao Zhang; Xiaowei Huang; Wenjie Ruan Activation Gradient based Poisoned Sample Detection Against Backdoor Attacks. (31%)Danni Yuan; Shaokui Wei; Mingda Zhang; Li Liu; Baoyuan Wu Poisoned ChatGPT Finds Work for Idle Hands: Exploring Developers' Coding Practices with Insecure Suggestions from Poisoned AI Models. (22%)Sanghak Oh; Kiho Lee; Seonhye Park; Doowon Kim; Hyoungshick Kim Promoting Counterfactual Robustness through Diversity. (13%)Francesco Leofante; Nico Potyka Resilient Path Planning for UAVs in Data Collection under Adversarial Attacks. (10%)Xueyuan Wang; M. Cenk Gursoy Adversarial Camera Patch: An Effective and Robust Physical-World Attack on Object Detectors. (1%)Kalibinuer Tiliwalidi Robust Graph Neural Network based on Graph Denoising. (1%)Victor M. Tenorio; Samuel Rey; Antonio G. Marques 2023-12-10 Data-Free Hard-Label Robustness Stealing Attack. (86%)Xiaojian Yuan; Kejiang Chen; Wen Huang; Jie Zhang; Weiming Zhang; Nenghai Yu A Practical Survey on Emerging Threats from AI-driven Voice Attacks: How Vulnerable are Commercial Voice Control Systems? (76%)Yuanda Wang; Qiben Yan; Nikolay Ivanov; Xun Chen An Ambiguity Measure for Recognizing the Unknowns in Deep Learning. (12%)Roozbeh Yousefzadeh METAL: Metamorphic Testing Framework for Analyzing Large-Language Model Qualities. (2%)Sangwon Hyun; Mingyu Guo; M. Ali Babar 2023-12-09 Poisoning $\times$ Evasion: Symbiotic Adversarial Robustness for Graph Neural Networks. (99%)Ege Erdogan; Simon Geisler; Stephan Günnemann Improving Adversarial Robust Fairness via Anti-Bias Soft Label Distillation. (98%)Shiji Zhao; Xizhe Wang; Xingxing Wei Dynamic Adversarial Attacks on Autonomous Driving Systems. (98%)Amirhosein Chahe; Chenan Wang; Abhishek Jeyapratap; Kaidi Xu; Lifeng Zhou Initialization Matters for Adversarial Transfer Learning. (76%)Andong Hua; Jindong Gu; Zhiyu Xue; Nicholas Carlini; Eric Wong; Yao Qin 2023-12-08 HC-Ref: Hierarchical Constrained Refinement for Robust Adversarial Training of GNNs. (99%)Xiaobing Pei; Haoran Yang; Gang Shen SA-Attack: Improving Adversarial Transferability of Vision-Language Pre-training Models via Self-Augmentation. (99%)Bangyan He; Xiaojun Jia; Siyuan Liang; Tianrui Lou; Yang Liu; Xiaochun Cao MIMIR: Masked Image Modeling for Mutual Information-based Adversarial Robustness. (99%)Xiaoyun Xu; Shujian Yu; Jingzheng Wu; Stjepan Picek BELT: Old-School Backdoor Attacks can Evade the State-of-the-Art Defense with Backdoor Exclusivity Lifting. (96%)Huming Qiu; Junjie Sun; Mi Zhang; Xudong Pan; Min Yang An adversarial attack approach for eXplainable AI evaluation on deepfake detection models. (38%)Balachandar Gowrisankar; Vrizlynn L. L. Thing A Red Teaming Framework for Securing AI in Maritime Autonomous Systems. (3%)Mathew J. Walter; Aaron Barrett; Kimberly Tam Annotation-Free Group Robustness via Loss-Based Resampling. (2%)Mahdi Ghaznavi; Hesam Asadollahzadeh; HamidReza Yaghoubi Araghi; Fahimeh Hosseini Noohdani; Mohammad Hossein Rohban; Mahdieh Soleymani Baghshah HuRef: HUman-REadable Fingerprint for Large Language Models. (1%)Boyi Zeng; Chenghu Zhou; Xinbing Wang; Zhouhan Lin 2023-12-07 MimicDiffusion: Purifying Adversarial Perturbation via Mimicking Clean Diffusion Model. (99%)Kaiyu Song; Hanjiang Lai OT-Attack: Enhancing Adversarial Transferability of Vision-Language Models via Optimal Transport Optimization. (99%)Dongchen Han; Xiaojun Jia; Yang Bai; Jindong Gu; Yang Liu; Xiaochun Cao Diffence: Fencing Membership Privacy With Diffusion Models. (97%)Yuefeng Peng; Ali Naseh; Amir Houmansadr FreqFed: A Frequency Analysis-Based Approach for Mitigating Poisoning Attacks in Federated Learning. (70%)Hossein Fereidooni; Alessandro Pegoraro; Phillip Rieger; Alexandra Dmitrienko; Ahmad-Reza Sadeghi Forcing Generative Models to Degenerate Ones: The Power of Data Poisoning Attacks. (64%)Shuli Jiang; Swanand Ravindra Kadhe; Yi Zhou; Ling Cai; Nathalie Baracaldo DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions. (15%)Fangzhou Wu; Xiaogeng Liu; Chaowei Xiao 2023-12-06 Defense against ML-based Power Side-channel Attacks on DNN Accelerators with Adversarial Attacks. (98%)Xiaobei Yan; Chip Hong Chang; Tianwei Zhang Defense Against Adversarial Attacks using Convolutional Auto-Encoders. (97%)Shreyasi Mandal Node-aware Bi-smoothing: Certified Robustness against Graph Injection Attacks. (88%)Yuni Lai; Yulin Zhu; Bailin Pan; Kai Zhou RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training. (54%)Jaehyung Kim; Yuning Mao; Rui Hou; Hanchao Yu; Davis Liang; Pascale Fung; Qifan Wang; Fuli Feng; Lifu Huang; Madian Khabsa Detecting Voice Cloning Attacks via Timbre Watermarking. (13%)Chang Liu; Jie Zhang; Tianwei Zhang; Xi Yang; Weiming Zhang; Nenghai Yu Synthesizing Physical Backdoor Datasets: An Automated Framework Leveraging Deep Generative Models. (11%)Sze Jue Yang; Chinh D. La; Quang H. Nguyen; Eugene Bagdasaryan; Kok-Seng Wong; Anh Tuan Tran; Chee Seng Chan; Khoa D. Doan Dr. Jekyll and Mr. Hyde: Two Faces of LLMs. (4%)Matteo Gioele Collu; Tom Janssen-Groesbeek; Stefanos Koffas; Mauro Conti; Stjepan Picek MICRO: Model-Based Offline Reinforcement Learning with a Conservative Bellman Operator. (2%)Xiao-Yin Liu; Xiao-Hu Zhou; Guo-Tao Li; Hao Li; Mei-Jiang Gui; Tian-Yu Xiang; De-Xing Huang; Zeng-Guang Hou 2023-12-05 Generating Visually Realistic Adversarial Patch. (99%)Xiaosen Wang; Kunyu Wang A Simple Framework to Enhance the Adversarial Robustness of Deep Learning-based Intrusion Detection System. (99%)Xinwei Yuan; Shu Han; Wei Huang; Hongliang Ye; Xianglong Kong; Fan Zhang Realistic Scatterer Based Adversarial Attacks on SAR Image Classifiers. (99%)Tian Ye; Rajgopal Kannan; Viktor Prasanna; Carl Busart; Lance Kaplan ScAR: Scaling Adversarial Robustness for LiDAR Object Detection. (99%)Xiaohu Lu; Hayder Radha Class Incremental Learning for Adversarial Robustness. (98%)Seungju Cho; Hongsin Lee; Changick Kim Provable Adversarial Robustness for Group Equivariant Tasks: Graphs, Point Clouds, Molecules, and More. (89%)Jan Schuchardt; Yan Scholten; Stephan Günnemann On the Robustness of Large Multimodal Models Against Image Adversarial Attacks. (69%)Xuanimng Cui; Alejandro Aparcedo; Young Kyun Jang; Ser-Nam Lim Scaling Laws for Adversarial Attacks on Language Model Activations. (50%)Stanislav Fort Indirect Gradient Matching for Adversarial Robust Distillation. (13%)Hongsin Lee; Seungju Cho; Changick Kim Robust Backdoor Detection for Deep Learning via Topological Evolution Dynamics. (3%)Xiaoxing Mo; Yechao Zhang; Leo Yu Zhang; Wei Luo; Nan Sun; Shengshan Hu; Shang Gao; Yang Xiang Prompt Optimization via Adversarial In-Context Learning. (3%)Xuan Long Do; Yiran Zhao; Hannah Brown; Yuxi Xie; James Xu Zhao; Nancy F. Chen; Kenji Kawaguchi; Michael Qizhe Xie; Junxian He Privacy-Preserving Task-Oriented Semantic Communications Against Model Inversion Attacks. (2%)Yanhu Wang; Shuaishuai Guo; Yiqin Deng; Haixia Zhang; Yuguang Fang Machine Vision Therapy: Multimodal Large Language Models Can Enhance Visual Robustness via Denoising In-Context Learning. (2%)Zhuo Huang; Chang Liu; Yinpeng Dong; Hang Su; Shibao Zheng; Tongliang Liu 2023-12-04 Adversarial Medical Image with Hierarchical Feature Hiding. (99%)Qingsong Yao; Zecheng He; Yuexiang Li; Yi Lin; Kai Ma; Yefeng Zheng; S. Kevin Zhou InstructTA: Instruction-Tuned Targeted Attack for Large Vision-Language Models. (99%)Xunguang Wang; Zhenlan Ji; Pingchuan Ma; Zongjie Li; Shuai Wang Singular Regularization with Information Bottleneck Improves Model's Adversarial Robustness. (98%)Guanlin Li; Naishan Zheng; Man Zhou; Jie Zhang; Tianwei Zhang Two-stage optimized unified adversarial patch for attacking visible-infrared cross-modal detectors in the physical world. (12%)Chengyin Hu; Weiwen Shi Auto DP-SGD: Dual Improvements of Privacy and Accuracy via Automatic Clipping Threshold and Noise Multiplier Estimation. (1%)Sai Venkatesh Chilukoti; Md Imran Hossen; Liqun Shan; Vijay Srinivas Tida; Xiai Hei Rejuvenating image-GPT as Strong Visual Representation Learners. (1%)Sucheng Ren; Zeyu Wang; Hongru Zhu; Junfei Xiao; Alan Yuille; Cihang Xie 2023-12-03 QuantAttack: Exploiting Dynamic Quantization to Attack Vision Transformers. (99%)Amit Baras; Alon Zolfi; Yuval Elovici; Asaf Shabtai OCGEC: One-class Graph Embedding Classification for DNN Backdoor Detection. (61%)Haoyu Jiang; Haiyang Yu; Nan Li; Ping Yi Evaluating the Security of Satellite Systems. (16%)Roy Peled; Eran Aizikovich; Edan Habler; Yuval Elovici; Asaf Shabtai Exploring Adversarial Robustness of LiDAR-Camera Fusion Model in Autonomous Driving. (13%)Bo Yang; Xiaoyu Ji; Xiaoyu Ji; Xiaoyu Ji; Xiaoyu Ji Towards Sample-specific Backdoor Attack with Clean Labels via Attribute Trigger. (2%)Yiming Li; Mingyan Zhu; Junfeng Guo; Tao Wei; Shu-Tao Xia; Zhan Qin 2023-12-02 TranSegPGD: Improving Transferability of Adversarial Examples on Semantic Segmentation. (99%)Xiaojun Jia; Jindong Gu; Yihao Huang; Simeng Qin; Qing Guo; Yang Liu; Xiaochun Cao Rethinking PGD Attack: Is Sign Function Necessary? (98%)Junjie Yang; Tianlong Chen; Xuxi Chen; Zhangyang Wang; Yingbin Liang PROFL: A Privacy-Preserving Federated Learning Method with Stringent Defense Against Poisoning Attacks. (61%)Yisheng Zhong; Li-Ping Wang Mendata: A Framework to Purify Manipulated Training Data. (2%)Zonghao Huang; Neil Gong; Michael K. Reiter 2023-12-01 PyraTrans: Learning Attention-Enriched Multi-Scale Pyramid Network from Pre-Trained Transformers for Effective Malicious URL Detection. (69%)Ruitong Liu; Yanbin Wang; Zhenhao Guo; Haitao Xu; Zhan Qin; Wenrui Ma; Fan Zhang Survey of Security Issues in Memristor-based Machine Learning Accelerators for RF Analysis. (22%)William Lillis; Max Cohen Hoffing; Wayne Burleson Deep Generative Attacks and Countermeasures for Data-Driven Offline Signature Verification. (10%)An Ngo; MinhPhuong Cao; Rajesh Kumar Temperature Balancing, Layer-wise Weight Analysis, and Neural Network Training. (1%)Yefan Zhou; Tianyu Pang; Keqin Liu; Charles H. Martin; Michael W. Mahoney; Yaoqing Yang Crystal: Enhancing Blockchain Mining Transparency with Quorum Certificate. (1%)Jianyu Niu; Fangyu Gai; Runchao Han; Ren Zhang; Yinqian Zhang; Chen Feng 2023-11-30 Improving the Robustness of Quantized Deep Neural Networks to White-Box Attacks using Stochastic Quantization and Information-Theoretic Ensemble Training. (98%)Saurabh Farkya; Aswin Raghavan; Avi Ziskind Adversarial Attacks and Defenses for Wireless Signal Classifiers using CDI-aware GANs. (98%)Sujata Sinha; Alkan Soysal Universal Backdoor Attacks. (97%)Benjamin Schneider; Nils Lukas; Florian Kerschbaum Fool the Hydra: Adversarial Attacks against Multi-view Object Detection Systems. (97%)Bilel Tarchoun; Quazi Mishkatul Alam; Nael Abu-Ghazaleh; Ihsen Alouani Corrupting Convolution-based Unlearnable Datasets with Pixel-based Image Transformations. (88%)Xianlong Wang; Shengshan Hu; Minghui Li; Zhifei Yu; Ziqi Zhou; Leo Yu Zhang; Hai Jin Optimal Attack and Defense for Reinforcement Learning. (76%)Jeremy McMahan; Young Wu; Xiaojin Zhu; Qiaomin Xie Can Protective Perturbation Safeguard Personal Data from Being Exploited by Stable Diffusion? (74%)Zhengyue Zhao; Jinhao Duan; Kaidi Xu; Chenan Wang; Rui Zhangp Zidong Dup Qi Guo; Xing Hu Improving Adversarial Transferability via Model Alignment. (68%)Avery Ma; Amir-massoud Farahmand; Yangchen Pan; Philip Torr; Jindong Gu Data-Agnostic Model Poisoning against Federated Learning: A Graph Autoencoder Approach. (62%)Kai Li; Jingjing Zheng; Xin Yuan; Wei Ni; Ozgur B. Akan; H. Vincent Poor Mark My Words: Analyzing and Evaluating Language Model Watermarks. (2%)Julien Piet; Chawin Sitawarin; Vivian Fang; Norman Mu; David Wagner 2023-11-29 Improving the Robustness of Transformer-based Large Language Models with Dynamic Attention. (98%)Lujia Shen; Yuwen Pu; Shouling Ji; Changjiang Li; Xuhong Zhang; Chunpeng Ge; Ting Wang Group-wise Sparse and Explainable Adversarial Attacks. (96%)Shpresim Sadiku; Moritz Wagner; Sebastian Pokutta Quantum Neural Networks under Depolarization Noise: Exploring White-Box Attacks and Defenses. (88%)David Winderl; Nicola Franco; Jeanette Miriam Lorenz On the Adversarial Robustness of Graph Contrastive Learning Methods. (83%)Filippo Guerranti; Zinuo Yi; Anna Starovoit; Rafiq Kamel; Simon Geisler; Stephan Günnemann Adversarial Robust Memory-Based Continual Learner. (81%)Xiaoyue Mi; Fan Tang; Zonghan Yang; Danding Wang; Juan Cao; Peng Li; Yang Liu Improving Faithfulness for Vision Transformers. (80%)Lijie Hu; Yixin Liu; Ninghao Liu; Mengdi Huai; Lichao Sun; Di Wang TARGET: Template-Transferable Backdoor Attack Against Prompt-based NLP Models via GPT4. (68%)Zihao Tan; Qingliang Chen; Yongjian Huang; Chen Liang Topology-Preserving Adversarial Training. (10%)Xiaoyue Mi; Fan Tang; Yepeng Weng; Danding Wang; Juan Cao; Sheng Tang; Peng Li; Yang Liu Query-Relevant Images Jailbreak Large Multi-Modal Models. (9%)Xin Liu; Yichen Zhu; Yunshi Lan; Chao Yang; Yu Qiao Analyzing and Explaining Image Classifiers via Diffusion Guidance. (8%)Maximilian Augustin; Yannic Neuhaus; Matthias Hein Poisoning Attacks Against Contrastive Recommender Systems. (2%)Zongwei Wang; Junliang Yu; Min Gao; Hongzhi Yin; Bin Cui; Shazia Sadiq SenTest: Evaluating Robustness of Sentence Encoders. (2%)Tanmay Chavan; Shantanu Patankar; Aditya Kane; Omkar Gokhale; Geetanjali Kale; Raviraj Joshi CLIPC8: Face liveness detection algorithm based on image-text pairs and contrastive learning. (1%)Xu Liu; Shu Zhou; Yurong Song; Wenzhe Luo; Xin Zhang Unveiling the Implicit Toxicity in Large Language Models. (1%)Jiaxin Wen; Pei Ke; Hao Sun; Zhexin Zhang; Chengfei Li; Jinfeng Bai; Minlie Huang 2023-11-28 Vulnerability Analysis of Transformer-based Optical Character Recognition to Adversarial Attacks. (99%)Lucas Beerens; Desmond J. Higham NeRFTAP: Enhancing Transferability of Adversarial Patches on Face Recognition using Neural Radiance Fields. (99%)Xiaoliang Liu; Furao Shen; Feng Han; Jian Zhao; Changhai Nie Efficient Key-Based Adversarial Defense for ImageNet by Using Pre-trained Model. (98%)AprilPyone MaungMaung; Isao Echizen; Hitoshi Kiya RADAP: A Robust and Adaptive Defense Against Diverse Adversarial Patches on Face Recognition. (92%)Xiaoliang Liu; Furao Shen; Jian Zhao; Changhai Nie STR-Cert: Robustness Certification for Deep Text Recognition on Deep Learning Pipelines and Vision Transformers. (26%)Daqian Shao; Lukas Fesser; Marta Kwiatkowska 1-Lipschitz Layers Compared: Memory, Speed, and Certifiable Robustness. (13%)Bernd Prach; Fabio Brau; Giorgio Buttazzo; Christoph H. Lampert Scalable Extraction of Training Data from (Production) Language Models. (10%)Milad Nasr; Nicholas Carlini; Jonathan Hayase; Matthew Jagielski; A. Feder Cooper; Daphne Ippolito; Christopher A. Choquette-Choo; Eric Wallace; Florian Tramèr; Katherine Lee Cooperative Abnormal Node Detection with Adversary Resistance: A Probabilistic Approach. (10%)Yingying Huangfu; Tian Bai On robust overfitting: adversarial training induced distribution matters. (1%)Runzhi Tian; Yongyi Mao Understanding the (Extra-)Ordinary: Validating Deep Model Decisions with Prototypical Concept-based Explanations. (1%)Maximilian Dreyer; Reduan Achtibat; Wojciech Samek; Sebastian Lapuschkin Shadows Don't Lie and Lines Can't Bend! Generative Models don't know Projective Geometry...for now. (1%)Ayush Sarkar; Hanlin Mai; Amitabh Mahapatra; Svetlana Lazebnik; D. A. Forsyth; Anand Bhattad 2023-11-27 RetouchUAA: Unconstrained Adversarial Attack via Image Retouching. (99%)Mengda Xie; Yiling He; Meie Fang Adversaral Doodles: Interpretable and Human-drawable Attacks Provide Describable Insights. (99%)Ryoya Nara; Yusuke Matsui Rethinking Mixup for Improving the Adversarial Transferability. (98%)Xiaosen Wang; Zeyuan Yin Instruct2Attack: Language-Guided Semantic Adversarial Attacks. (98%)Jiang Liu; Chen Wei; Yuxiang Guo; Heng Yu; Alan Yuille; Soheil Feizi; Chun Pong Lau; Rama Chellappa CLAP: Contrastive Learning with Augmented Prompts for Robustness on Pretrained Vision-Language Models. (95%)Yichao Cai; Yuhang Liu; Zhen Zhang; Javen Qinfeng Shi A Survey on Vulnerability of Federated Learning: A Learning Algorithm Perspective. (50%)Xianghua Xie; Chen Hu; Hanchi Ren; Jingjing Deng Threshold Breaker: Can Counter-Based RowHammer Prevention Mechanisms Truly Safeguard DRAM? (31%)Ranyang Zhou; Jacqueline Liu; Sabbir Ahmed; Nakul Kochar; Adnan Siraj Rakin; Shaahin Angizi Elijah: Eliminating Backdoors Injected in Diffusion Models via Distribution Shift. (31%)Shengwei An; Sheng-Yen Chou; Kaiyuan Zhang; Qiuling Xu; Guanhong Tao; Guangyu Shen; Siyuan Cheng; Shiqing Ma; Pin-Yu Chen; Tsung-Yi Ho; Xiangyu Zhang Distributed Attacks over Federated Reinforcement Learning-enabled Cell Sleep Control. (22%)Han Zhang; Hao Zhou; Medhat Elsayed; Majid Bavand; Raimundas Gaigalas; Yigit Ozcan; Melike Erol-Kantarci "Do Users fall for Real Adversarial Phishing?" Investigating the Human response to Evasive Webpages. (15%)Ajka Draganovic; Savino Dambra; Javier Aldana Iuit; Kevin Roundy; Giovanni Apruzzese How Many Unicorns Are in This Image? A Safety Evaluation Benchmark for Vision LLMs. (12%)Haoqin Tu; Chenhang Cui; Zijun Wang; Yiyang Zhou; Bingchen Zhao; Junlin Han; Wangchunshu Zhou; Huaxiu Yao; Cihang Xie Microarchitectural Security of AWS Firecracker VMM for Serverless Cloud Platforms. (1%)Zane Worcester Polytechnic Institute Weissman; Thore University of Lübeck Tiemann; Thomas University of Lübeck Eisenbarth; Berk Worcester Polytechnic Institute Sunar 2023-11-26 Adversarial Purification of Information Masking. (99%)Sitong Liu; Zhichao Lian; Shuangquan Zhang; Liang Xiao Having Second Thoughts? Let's hear it. (56%)Jung H. Lee; Sujith Vijayan BadCLIP: Trigger-Aware Prompt Learning for Backdoor Attacks on CLIP. (13%)Jiawang Bai; Kuofeng Gao; Shaobo Min; Shu-Tao Xia; Zhifeng Li; Wei Liu Confidence Is All You Need for MI Attacks. (2%)Abhishek Sinha; Himanshi Tibrewal; Mansi Gupta; Nikhar Waghela; Shivank Garg 2023-11-25 Mixing Classifiers to Alleviate the Accuracy-Robustness Trade-Off. (26%)Yatong Bai; Brendon G. Anderson; Somayeh Sojoudi Robust Graph Neural Networks via Unbiased Aggregation. (10%)Ruiqi Feng; Zhichao Hou; Tyler Derr; Xiaorui Liu Effective Backdoor Mitigation Depends on the Pre-training Objective. (10%)Sahil Verma; Gantavya Bhatt; Avi Schwarzschild; Soumye Singhal; Arnav Mohanty Das; Chirag Shah; John P Dickerson; Jeff Bilmes 2023-11-24 Trainwreck: A damaging adversarial attack on image classifiers. (99%)Jan Zahálka Segment (Almost) Nothing: Prompt-Agnostic Adversarial Attacks on Segmentation Models. (96%)Francesco Croce; Matthias Hein Universal Jailbreak Backdoors from Poisoned Human Feedback. (1%)Javier Rando; Florian Tramèr 2023-11-23 When Side-Channel Attacks Break the Black-Box Property of Embedded Artificial Intelligence. (99%)Benoit Coqueret; Mathieu Carbone; Olivier Sentieys; Gabriel Zaid Adversarial defense based on distribution transfer. (99%)Jiahao Chen; Diqun Yan; Li Dong Robust and Interpretable COVID-19 Diagnosis on Chest X-ray Images using Adversarial Training. (68%)Karina Yang; Alexis Bennett; Dominique Duncan Presentation Attack Detection using Convolutional Neural Networks and Local Binary Patterns. (1%)Justin Spencer; Deborah Lawrence; Prosenjit Chatterjee; Kaushik Roy; Albert Esterline; Jung-Hee Kim 2023-11-22 A Survey of Adversarial CAPTCHAs on its History, Classification and Generation. (99%)Zisheng Xu; Qiao Yan; F. Richard Yu; Victor C. M. Leung Transfer Attacks and Defenses for Large Language Models on Coding Tasks. (99%)Chi Zhang; Zifan Wang; Ravi Mangal; Matt Fredrikson; Limin Jia; Corina Pasareanu Panda or not Panda? Understanding Adversarial Attacks with Interactive Visualization. (98%)Yuzhe You; Jarvis Tse; Jian Zhao Hard Label Black Box Node Injection Attack on Graph Neural Networks. (93%)Yu Zhou; Zihao Dong; Guofeng Zhang; Jingchen Tang Security and Privacy Challenges in Deep Learning Models. (74%)Gopichandh Golla A Somewhat Robust Image Watermark against Diffusion-based Editing Models. (50%)Mingtian Tan; Tianhao Wang; Somesh Jha OASIS: Offsetting Active Reconstruction Attacks in Federated Learning. (2%)Tre' R. Jeter; Truc Nguyen; Raed Alharbi; My T. Thai 2023-11-21 SD-NAE: Generating Natural Adversarial Examples with Stable Diffusion. (96%)Yueqian Lin; Jingyang Zhang; Yiran Chen; Hai Li Stable Unlearnable Example: Enhancing the Robustness of Unlearnable Examples via Stable Error-Minimizing Noise. (96%)Yixin Liu; Kaidi Xu; Xun Chen; Lichao Sun Attention Deficit is Ordered! Fooling Deformable Vision Transformers with Collaborative Adversarial Patches. (75%)Quazi Mishkatul Alam; Bilel Tarchoun; Ihsen Alouani; Nael Abu-Ghazaleh Attacking Motion Planners Using Adversarial Perception Errors. (69%)Jonathan Sadeghi; Nicholas A. Lord; John Redford; Romain Mueller Toward Robust Imperceptible Perturbation against Unauthorized Text-to-image Diffusion-based Synthesis. (62%)Yixin Liu; Chenrui Fan; Yutong Dai; Xun Chen; Pan Zhou; Lichao Sun Iris Presentation Attack: Assessing the Impact of Combining Vanadium Dioxide Films with Artificial Eyes. (1%)Darshika Jauhari; Renu Sharma; Cunjian Chen; Nelson Sepulveda; Arun Ross 2023-11-20 ODDR: Outlier Detection & Dimension Reduction Based Defense Against Adversarial Patches. (99%)Nandish Chattopadhyay; Amira Guesmi; Muhammad Abdullah Hanif; Bassem Ouni; Muhammad Shafique DefensiveDR: Defending against Adversarial Patches using Dimensionality Reduction. (99%)Nandish Chattopadhyay; Amira Guesmi; Muhammad Abdullah Hanif; Bassem Ouni; Muhammad Shafique Generating Valid and Natural Adversarial Examples with Large Language Models. (99%)Zimu Wang; Wei Wang; Qi Chen; Qiufeng Wang; Anh Nguyen AdvGen: Physical Adversarial Attack on Face Presentation Attack Detection Systems. (99%)Sai Amrit Patnaik; Shivali Chansoriya; Anil K. Jain; Anoop M. Namboodiri Beyond Boundaries: A Comprehensive Survey of Transferable Attacks on AI Systems. (50%)Guangjing Wang; Ce Zhou; Yuanda Wang; Bocheng Chen; Hanqing Guo; Qiben Yan Understanding Variation in Subpopulation Susceptibility to Poisoning Attacks. (15%)Evan Rose; Fnu Suya; David Evans Training robust and generalizable quantum models. (10%)Julian Berberich; Daniel Fink; Daniel Pranjić; Christian Tutschku; Christian Holm BrainWash: A Poisoning Attack to Forget in Continual Learning. (4%)Ali Abbasi; Parsa Nooralinejad; Hamed Pirsiavash; Soheil Kolouri 2023-11-19 Adversarial Prompt Tuning for Vision-Language Models. (98%)Jiaming Zhang; Xingjun Ma; Xin Wang; Lingyu Qiu; Jiaqi Wang; Yu-Gang Jiang; Jitao Sang Token-Level Adversarial Prompt Detection Based on Perplexity Measures and Contextual Information. (78%)Zhengmian Hu; Gang Wu; Saayan Mitra; Ruiyi Zhang; Tong Sun; Heng Huang; Viswanathan Swaminathan BadCLIP: Dual-Embedding Guided Backdoor Attack on Multimodal Contrastive Learning. (69%)Siyuan Liang; Mingli Zhu; Aishan Liu; Baoyuan Wu; Xiaochun Cao; Ee-Chien Chang EditShield: Protecting Unauthorized Image Editing by Instruction-guided Diffusion Models. (10%)Ruoxi Chen; Haibo Jin; Jinyin Chen; Lichao Sun 2023-11-18 Boost Adversarial Transferability by Uniform Scale and Mix Mask Method. (99%)Tao Wang; Zijian Ying; Qianmu Li; zhichao Lian Improving Adversarial Transferability by Stable Diffusion. (99%)Jiayang Liu; Siyu Zhu; Siyuan Liang; Jie Zhang; Han Fang; Weiming Zhang; Ee-Chien Chang Attention-Based Real-Time Defenses for Physical Adversarial Attacks in Vision Applications. (92%)Giulio Rossolini; Alessandro Biondi; Giorgio Buttazzo TextGuard: Provable Defense against Backdoor Attacks on Text Classification. (82%)Hengzhi Pei; Jinyuan Jia; Wenbo Guo; Bo Li; Dawn Song Robust Network Slicing: Multi-Agent Policies, Adversarial Attacks, and Defensive Strategies. (1%)Feng Wang; M. Cenk Gursoy; Senem Velipasalar 2023-11-17 Breaking Temporal Consistency: Generating Video Universal Adversarial Perturbations Using Image Models. (97%)Hee-Seon Kim; Minji Son; Minbeom Kim; Myung-Joon Kwon; Changick Kim PACOL: Poisoning Attacks Against Continual Learners. (93%)Huayu Li; Gregory Ditzler Two-Factor Authentication Approach Based on Behavior Patterns for Defeating Puppet Attacks. (1%)Wenhao Wang; Guyue Li; Zhiming Chu; Haobo Li; Daniele Faccio 2023-11-16 Breaking Boundaries: Balancing Performance and Robustness in Deep Wireless Traffic Forecasting. (99%)Romain Ilbert; Thai V. Hoang; Zonghua Zhang; Themis Palpanas Hijacking Large Language Models via Adversarial In-Context Learning. (75%)Yao Qiang; Xiangyu Zhou; Dongxiao Zhu Cognitive Overload: Jailbreaking Large Language Models with Overloaded Logical Thinking. (54%)Nan Xu; Fei Wang; Ben Zhou; Bang Zheng Li; Chaowei Xiao; Muhao Chen Test-time Backdoor Mitigation for Black-Box Large Language Models with Defensive Demonstrations. (38%)Wenjie Mo; Jiashu Xu; Qin Liu; Jiongxiao Wang; Jun Yan; Chaowei Xiao; Muhao Chen On the Exploitability of Reinforcement Learning with Human Feedback for Large Language Models. (13%)Jiongxiao Wang; Junlin Wu; Muhao Chen; Yevgeniy Vorobeychik; Chaowei Xiao Towards Improving Robustness Against Common Corruptions using Mixture of Class Specific Experts. (2%)Shashank Kotyan; Danilo Vasconcellos Vargas Understanding the Effectiveness of Large Language Models in Detecting Security Vulnerabilities. (2%)Avishree Khare; Saikat Dutta; Ziyang Li; Alaia Solko-Breslin; Rajeev Alur; Mayur Naik Bergeron: Combating Adversarial Attacks through a Conscience-Based Alignment Framework. (2%)Matthew Pisano; Peter Ly; Abraham Sanders; Bingsheng Yao; Dakuo Wang; Tomek Strzalkowski; Mei Si Towards more Practical Threat Models in Artificial Intelligence Security. (2%)Kathrin Grosse; Lukas Bieringer; Tarek Richard Besold; Alexandre Alahi You Cannot Escape Me: Detecting Evasions of SIEM Rules in Enterprise Networks. (1%)Rafael Uetz; Marco Herzog; Louis Hackländer; Simon Schwarz; Martin Henze 2023-11-15 Jailbreaking GPT-4V via Self-Adversarial Attacks with System Prompts. (99%)Yuanwei Wu; Xiang Li; Yixin Liu; Pan Zhou; Lichao Sun Backdoor Activation Attack: Attack Large Language Models using Activation Steering for Safety-Alignment. (74%)Haoran Wang; Kai Shu Fast Certification of Vision-Language Models Using Incremental Randomized Smoothing. (64%)A K Iowa State University Nirala; A New York University Joshi; C New York University Hegde; S Iowa State University Sarkar Adversarially Robust Spiking Neural Networks Through Conversion. (61%)Ozan Özdenizci; Robert Legenstein How Trustworthy are Open-Source LLMs? An Assessment under Malicious Demonstrations Shows their Vulnerabilities. (16%)Lingbo Mo; Boshi Wang; Muhao Chen; Huan Sun Defending Large Language Models Against Jailbreaking Attacks Through Goal Prioritization. (15%)Zhexin Zhang; Junxiao Yang; Pei Ke; Minlie Huang Privacy Threats in Stable Diffusion Models. (13%)Thomas Cilloni; Charles Fleming; Charles Walter MirrorNet: A TEE-Friendly Framework for Secure On-device DNN Inference. (2%)Ziyu Liu; Yukui Luo; Shijin Duan; Tong Zhou; Xiaolin Xu JAB: Joint Adversarial Prompting and Belief Augmentation. (1%)Ninareh Mehrabi; Palash Goyal; Anil Ramakrishna; Jwala Dhamala; Shalini Ghosh; Richard Zemel; Kai-Wei Chang; Aram Galstyan; Rahul Gupta Beyond Detection: Unveiling Fairness Vulnerabilities in Abusive Language Models. (1%)Yueqing Liang; Lu Cheng; Ali Payani; Kai Shu 2023-11-14 Towards Improving Robustness Against Common Corruptions in Object Detectors Using Adversarial Contrastive Learning. (99%)Shashank Kotyan; Danilo Vasconcellos Vargas Physical Adversarial Examples for Multi-Camera Systems. (99%)Ana Răduţoiu; Jan-Philipp Schulze; Philip Sperl; Konstantin Böttinger DALA: A Distribution-Aware LoRA-Based Adversarial Attack against Language Models. (99%)Yibo Wang; Xiangjue Dong; James Caverlee; Philip S. Yu On The Relationship Between Universal Adversarial Attacks And Sparse Representations. (98%)Dana Weitzner; Raja Giryes A Wolf in Sheep's Clothing: Generalized Nested Jailbreak Prompts can Fool Large Language Models Easily. (62%)Peng Ding; Jun Kuang; Dan Ma; Xuezhi Cao; Yunsen Xian; Jiajun Chen; Shujian Huang Multi-Set Inoculation: Assessing Model Robustness Across Multiple Challenge Sets. (13%)Vatsal Gupta; Pranshu Pandya; Tushar Kataria; Vivek Gupta; Dan Roth The Perception-Robustness Tradeoff in Deterministic Image Restoration. (1%)Guy Ohayon; Tomer Michaeli; Michael Elad 2023-11-13 Adversarial Purification for Data-Driven Power System Event Classifiers with Diffusion Models. (99%)Yuanbin Cheng; Koji Yamashita; Jim Follum; Nanpeng Yu Parrot-Trained Adversarial Examples: Pushing the Practicality of Black-Box Audio Attacks against Speaker Recognition Models. (99%)Rui Duan; Zhe Qu; Leah Ding; Yao Liu; Zhuo Lu An Extensive Study on Adversarial Attack against Pre-trained Models of Code. (99%)Xiaohu Du; Ming Wen; Zichao Wei; Shangwen Wang; Hai Jin Untargeted Black-box Attacks for Social Recommendations. (96%)Wenqi Fan; Shijie Wang; Xiao-yong Wei; Xiaowei Mei; Qing Li On the Robustness of Neural Collapse and the Neural Collapse of Robustness. (80%)Jingtong Su; Ya Shi Zhang; Nikolaos Tsilivis; Julia Kempe Tabdoor: Backdoor Vulnerabilities in Transformer-based Neural Networks for Tabular Data. (64%)Bart Pleiter; Behrad Tajalli; Stefanos Koffas; Gorka Abad; Jing Xu; Martha Larson; Stjepan Picek 2023-11-12 Learning Globally Optimized Language Structure via Adversarial Training. (83%)Xuwang Yin Contractive Systems Improve Graph Neural Networks Against Adversarial Attacks. (70%)Moshe Eliasof; Davide Murari; Ferdia Sherry; Carola-Bibiane Schönlieb Analytical Verification of Deep Neural Network Performance for Time-Synchronized Distribution System State Estimation. (5%)Behrouz Azimian; Shiva Moshtagh; Anamitra Pal; Shanshan Ma DialMAT: Dialogue-Enabled Transformer with Moment-Based Adversarial Training. (1%)Kanta Kaneda; Ryosuke Korekata; Yuiga Wada; Shunya Nagashima; Motonari Kambara; Yui Iioka; Haruka Matsuo; Yuto Imai; Takayuki Nishimura; Komei Sugiura 2023-11-10 Flatness-aware Adversarial Attack. (99%)Mingyuan Fan; Xiaodan Li; Cen Chen; Yinggui Wang Robust Adversarial Attacks Detection for Deep Learning based Relative Pose Estimation for Space Rendezvous. (99%)Ziwei Wang; Nabil Aouf; Jose Pizarro; Christophe Honvault Fight Fire with Fire: Combating Adversarial Patch Attacks using Pattern-randomized Defensive Patches. (98%)Jianan Feng; Jiachun Li; Changqing Miao; Jianjun Huang; Wei You; Wenchang Shi; Bin Liang Resilient and constrained consensus against adversarial attacks: A distributed MPC framework. (84%)Henglai Wei; Kunwu Zhang; Hui Zhang; Yang Shi CALLOC: Curriculum Adversarial Learning for Secure and Robust Indoor Localization. (1%)Danish Gufran; Sudeep Pasricha Practical Membership Inference Attacks against Fine-tuned Large Language Models via Self-prompt Calibration. (1%)Wenjie Fu; Huandong Wang; Chen Gao; Guanghua Liu; Yong Li; Tao Jiang 2023-11-09 ABIGX: A Unified Framework for eXplainable Fault Detection and Classification. (68%)Yue Zhuo; Jinchuan Qian; Zhihuan Song; Zhiqiang Ge Honest Score Client Selection Scheme: Preventing Federated Learning Label Flipping Attacks in Non-IID Scenarios. (50%)Yanli Li; Huaming Chen; Wei Bao; Zhengmeng Xu; Dong Yuan Scale-MIA: A Scalable Model Inversion Attack against Secure Federated Learning via Latent Space Reconstruction. (15%)Shanghao Shi; Ning Wang; Yang Xiao; Chaoyu Zhang; Yi Shi; Y. Thomas Hou; Wenjing Lou FigStep: Jailbreaking Large Vision-language Models via Typographic Visual Prompts. (1%)Yichen Gong; Delong Ran; Jinyuan Liu; Conglei Wang; Tianshuo Cong; Anyu Wang; Sisi Duan; Xiaoyun Wang FireMatch: A Semi-Supervised Video Fire Detection Network Based on Consistency and Distribution Alignment. (1%)Qinghua Lin; Zuoyong Li; Kun Zeng; Haoyi Fan; Wei Li; Xiaoguang Zhou 2023-11-08 Constrained Adaptive Attacks: Realistic Evaluation of Adversarial Examples and Robust Training of Deep Neural Networks for Tabular Data. (99%)Thibault Simonetto; Salah Ghamizi; Antoine Desjardins; Maxime Cordy; Yves Le Traon Army of Thieves: Enhancing Black-Box Model Extraction via Ensemble based sample selection. (70%)Akshit Jindal; Vikram Goyal; Saket Anand; Chetan Arora Frontier Language Models are not Robust to Adversarial Arithmetic, or "What do I need to say so you agree 2+2=5? (61%)C. Daniel Freeman; Laura Culp; Aaron Parisi; Maxwell L Bileschi; Gamaleldin F Elsayed; Alex Rizkowsky; Isabelle Simpson; Alex Alemi; Azade Nova; Ben Adlam; Bernd Bohnet; Gaurav Mishra; Hanie Sedghi; Igor Mordatch; Izzeddin Gur; Jaehoon Lee; JD Co-Reyes; Jeffrey Pennington; Kelvin Xu; Kevin Swersky; Kshiteej Mahajan; Lechao Xiao; Rosanne Liu; Simon Kornblith; Noah Constant; Peter J. Liu; Roman Novak; Yundi Qian; Noah Fiedel; Jascha Sohl-Dickstein SCAAT: Improving Neural Network Interpretability via Saliency Constrained Adaptive Adversarial Training. (10%)Rui Xu; Wenkang Qin; Peixiang Huang; Haowang; Lin Luo Domain Adaptive Object Detection via Balancing Between Self-Training and Adversarial Learning. (1%)Muhammad Akhtar Munir; Muhammad Haris Khan; M. Saquib Sarfraz; Mohsen Ali Counter-Empirical Attacking based on Adversarial Reinforcement Learning for Time-Relevant Scoring System. (1%)Xiangguo Sun; Hong Cheng; Hang Dong; Bo Qiao; Si Qin; Qingwei Lin 2023-11-07 Unveiling Safety Vulnerabilities of Large Language Models. (61%)George Kour; Marcel Zalmanovici; Naama Zwerdling; Esther Goldbraich; Ora Nova Fandina; Ateret Anaby-Tavor; Orna Raz; Eitan Farchi When Fairness Meets Privacy: Exploring Privacy Threats in Fair Binary Classifiers through Membership Inference Attacks. (10%)Huan Tian; Guangsheng Zhang; Bo Liu; Tianqing Zhu; Ming Ding; Wanlei Zhou Identifying and Mitigating Vulnerabilities in LLM-Integrated Applications. (2%)Fengqing Jiang; Zhangchen Xu; Luyao Niu; Boxin Wang; Jinyuan Jia; Bo Li; Radha Poovendran SoK: Security Below the OS -- A Security Analysis of UEFI. (1%)Priyanka Prakash Surve; Oleg Brodt; Mark Yampolskiy; Yuval Elovici; Asaf Shabtai Do LLMs exhibit human-like response biases? A case study in survey design. (1%)Lindia Tjuatja; Valerie Chen; Sherry Tongshuang Wu; Ameet Talwalkar; Graham Neubig 2023-11-06 Measuring Adversarial Datasets. (92%)Yuanchen Bai; Raoyi Huang; Vijay Viswanathan; Tzu-Sheng Kuo; Tongshuang Wu Can LLMs Follow Simple Rules? (68%)Norman Mu; Sarah Chen; Zifan Wang; Sizhe Chen; David Karamardian; Lulwa Aljeraisy; Basel Alomair; Dan Hendrycks; David Wagner Preserving Privacy in GANs Against Membership Inference Attack. (33%)Mohammadhadi Shateri; Francisco Messina; Fabrice Labeau; Pablo Piantanida Cal-DETR: Calibrated Detection Transformer. (4%)Muhammad Akhtar Munir; Salman Khan; Muhammad Haris Khan; Mohsen Ali; Fahad Shahbaz Khan 2023-11-05 ELEGANT: Certified Defense on the Fairness of Graph Neural Networks. (10%)Yushun Dong; Binchi Zhang; Hanghang Tong; Jundong Li 2023-11-04 From Trojan Horses to Castle Walls: Unveiling Bilateral Backdoor Effects in Diffusion Models. (22%)Zhuoshi Pan; Yuguang Yao; Gaowen Liu; Bingquan Shen; H. Vicky Zhao; Ramana Rao Kompella; Sijia Liu 2023-11-03 Efficient Black-Box Adversarial Attacks on Neural Text Detectors. (22%)Vitalii Fishchuk; Daniel Braun The Alignment Problem in Context. (2%)Raphaël Millière 2023-11-02 Adversary ML Resilience in Autonomous Driving Through Human Centered Perception Mechanisms. (99%)Aakriti Shah Towards Evaluating Transfer-based Attacks Systematically, Practically, and Fairly. (99%)Qizhang Li; Yiwen Guo; Wangmeng Zuo; Hao Chen Tensor Trust: Interpretable Prompt Injection Attacks from an Online Game. (93%)Sam Toyer; Olivia Watkins; Ethan Adrian Mendes; Justin Svegliato; Luke Bailey; Tiffany Wang; Isaac Ong; Karim Elmaaroufi; Pieter Abbeel; Trevor Darrell; Alan Ritter; Stuart Russell On the Lipschitz constant of random neural networks. (92%)Paul Geuchen; Thomas Heindl; Dominik Stöger; Felix Voigtlaender Universal Perturbation-based Secret Key-Controlled Data Hiding. (80%)Donghua Wang; Wen Yao; Tingsong Jiang; Xiaoqian Chen Distilling Out-of-Distribution Robustness from Vision-Language Foundation Models. (76%)Andy Zhou; Jindong Wang; Yu-Xiong Wang; Haohan Wang Assist Is Just as Important as the Goal: Image Resurfacing to Aid Model's Robust Prediction. (13%)Abhijith Sharma; Phil Munz; Apurva Narayan Robust Adversarial Reinforcement Learning via Bounded Rationality Curricula. (12%)Aryaman Reddi; Maximilian Tölle; Jan Peters; Georgia Chalvatzaki; Carlo D'Eramo Sequential Subset Matching for Dataset Distillation. (1%)Jiawei Du; Qin Shi; Joey Tianyi Zhou E(2) Equivariant Neural Networks for Robust Galaxy Morphology Classification. (1%)Sneh Pandya; Purvik Patel; Franc O; Jonathan Blazek Robust Identity Perceptual Watermark Against Deepfake Face Swapping. (1%)Tianyi Wang; Mengxiao Huang; Harry Cheng; Bin Ma; Yinglong Wang 2023-11-01 NEO-KD: Knowledge-Distillation-Based Adversarial Training for Robust Multi-Exit Neural Networks. (99%)Seokil Ham; Jungwuk Park; Dong-Jun Han; Jaekyun Moon Adversarial Examples in the Physical World: A Survey. (98%)Jiakai Wang; Donghua Wang; Jin Hu; Siyang Wu; Tingsong Jiang; Wen Yao; Aishan Liu; Xianglong Liu Optimal Cost Constrained Adversarial Attacks For Multiple Agent Systems. (80%)Ziqing Lu; Guanlin Liu; Lifeng Cai; Weiyu Xu Improving Robustness for Vision Transformer with a Simple Dynamic Scanning Augmentation. (76%)Shashank Kotyan; Danilo Vasconcellos Vargas MIST: Defending Against Membership Inference Attacks Through Membership-Invariant Subspace Training. (75%)Jiacheng Li; Ninghui Li; Bruno Ribeiro Robustness Tests for Automatic Machine Translation Metrics with Adversarial Attacks. (1%)Yichen Huang; Timothy Baldwin Open-Set Face Recognition with Maximal Entropy and Objectosphere Loss. (1%)Rafael Henrique Vareto; Yu Linghu; Terrance E. Boult; William Robson Schwartz; Manuel Günther 2023-10-31 Amoeba: Circumventing ML-supported Network Censorship via Adversarial Reinforcement Learning. (99%)Haoyu Liu; Alec F. Diallo; Paul Patras Robust Safety Classifier for Large Language Models: Adversarial Prompt Shield. (99%)Jinhwa Kim; Ali Derakhshan; Ian G. Harris LFAA: Crafting Transferable Targeted Adversarial Examples with Low-Frequency Perturbations. (99%)Kunyu Wang; Juluan Shi; Wenxuan Wang Magmaw: Modality-Agnostic Adversarial Attacks on Machine Learning-Based Wireless Communication Systems. (98%)Jung-Woo Chang; Ke Sun; Nasimeh Heydaribeni; Seira Hidano; Xinyu Zhang; Farinaz Koushanfar Is Robustness Transferable across Languages in Multilingual Neural Machine Translation? (26%)Leiyu Pan; Supryadi; Deyi Xiong Dynamic Batch Norm Statistics Update for Natural Robustness. (22%)Shahbaz Rezaei; Mohammad Sadegh Norouzzadeh In Search of Lost Online Test-time Adaptation: A Survey. (1%)Zixin Wang; Yadan Luo; Liang Zheng; Zhuoxiao Chen; Sen Wang; Zi Huang 2023-10-30 Label-Only Model Inversion Attacks via Knowledge Transfer. (83%)Ngoc-Bao Nguyen; Keshigeyan Chandrasegaran; Milad Abdollahzadeh; Ngai-Man Cheung Exploring Geometry of Blind Spots in Vision Models. (83%)Sriram Balasubramanian; Gaurang Sriramanan; Vinu Sankar Sadasivan; Soheil Feizi Adversarial Attacks and Defenses in Large Language Models: Old and New Threats. (74%)Leo Schwinn; David Dobre; Stephan Günnemann; Gauthier Gidel Generated Distributions Are All You Need for Membership Inference Attacks Against Generative Models. (61%)Minxing Zhang; Ning Yu; Rui Wen; Michael Backes; Yang Zhang Causal Fair Metric: Bridging Causality, Individual Fairness, and Adversarial Robustness. (33%)Ahmad-Reza Ehyaei; Golnoosh Farnadi; Samira Samadi Differentially Private Reward Estimation with Preference Feedback. (16%)Sayak Ray Chowdhury; Xingyu Zhou; Nagarajan Natarajan Asymmetric Diffusion Based Channel-Adaptive Secure Wireless Semantic Communications. (10%)Xintian Ren; Jun Wu; Hansong Xu; Qianqian Pan Privacy-Preserving Federated Learning over Vertically and Horizontally Partitioned Data for Financial Anomaly Detection. (1%)Swanand Ravindra Kadhe; Heiko Ludwig; Nathalie Baracaldo; Alan King; Yi Zhou; Keith Houck; Ambrish Rawat; Mark Purcell; Naoise Holohan; Mikio Takeuchi; Ryo Kawahara; Nir Drucker; Hayim Shaul; Eyal Kushnir; Omri Soceanu 2023-10-29 Blacksmith: Fast Adversarial Training of Vision Transformers via a Mixture of Single-step and Multi-step Methods. (99%)Mahdi Salmani; Alireza Dehghanpour Farashah; Mohammad Azizmalayeri; Mahdi Amiri; Navid Eslami; Mohammad Taghi Manzuri; Mohammad Hossein Rohban Boosting Decision-Based Black-Box Adversarial Attack with Gradient Priors. (98%)Han Liu; Xingshuo Huang; Xiaotong Zhang; Qimai Li; Fenglong Ma; Wei Wang; Hongyang Chen; Hong Yu; Xianchao Zhang BERT Lost Patience Won't Be Robust to Adversarial Slowdown. (98%)Zachary Coalson; Gabriel Ritter; Rakesh Bobba; Sanghyun Hong Adversarial Examples Are Not Real Features. (98%)Ang Li; Yifei Wang; Yiwen Guo; Yisen Wang IMPRESS: Evaluating the Resilience of Imperceptible Perturbations Against Unauthorized Data Usage in Diffusion-Based Generative AI. (82%)Bochuan Cao; Changjiang Li; Ting Wang; Jinyuan Jia; Bo Li; Jinghui Chen Poisoning Retrieval Corpora by Injecting Adversarial Passages. (68%)Zexuan Zhong; Ziqing Huang; Alexander Wettig; Danqi Chen Label Poisoning is All You Need. (54%)Rishi D. Jha; Jonathan Hayase; Sewoong Oh Robustifying Language Models with Test-Time Adaptation. (47%)Noah Thomas McDermott; Junfeng Yang; Chengzhi Mao Path Analysis for Effective Fault Localization in Deep Neural Networks. (1%)Soroush Hashemifar; Saeed Parsa; Akram Kalaee From Chatbots to PhishBots? -- Preventing Phishing scams created using ChatGPT, Google Bard and Claude. (1%)Sayak Saha Roy; Poojitha Thota; Krishna Vamsi Naragam; Shirin Nilizadeh 2023-10-28 Assessing and Improving Syntactic Adversarial Robustness of Pre-trained Models for Code Translation. (92%)Guang Yang; Yu Zhou; Xiangyu Zhang; Xiang Chen; Tingting Han; Taolue Chen Benchmark Generation Framework with Customizable Distortions for Image Classifier Robustness. (86%)Soumyendu Sarkar; Ashwin Ramesh Babu; Sajad Mousavi; Zachariah Carmichael; Vineet Gundecha; Sahand Ghorbanpour; Ricardo Luna; Gutierrez Antonio Guillen; Avisek Naug Purify++: Improving Diffusion-Purification with Advanced Diffusion Models and Control of Randomness. (61%)Boya Zhang; Weijian Luo; Zhihua Zhang Large Language Models Are Better Adversaries: Exploring Generative Clean-Label Backdoor Attacks Against Text Classifiers. (47%)Wencong You; Zayd Hammoudeh; Daniel Lowd Where have you been? A Study of Privacy Risk for Point-of-Interest Recommendation. (8%)Kunlin Cai; Jinghuai Zhang; Will Shand; Zhiqing Hong; Guang Wang; Desheng Zhang; Jianfeng Chi; Yuan Tian 2023-10-27 DiffAttack: Evasion Attacks Against Diffusion-Based Adversarial Purification. (99%)Mintong Kang; Dawn Song; Bo Li Understanding and Improving Ensemble Adversarial Defense. (99%)Yian Deng; Tingting Mu LipSim: A Provably Robust Perceptual Similarity Metric. (45%)Sara Ghazanfari; Alexandre Araujo; Prashanth Krishnamurthy; Farshad Khorrami; Siddharth Garg Elevating Code-mixed Text Handling through Auditory Information of Words. (5%)Mamta; Zishan Ahmad; Asif Ekbal Understanding Parameter Saliency via Extreme Value Theory. (1%)Shuo Wang; Issei Sato 2023-10-26 Unscrambling the Rectification of Adversarial Attacks Transferability across Computer Networks. (99%)Ehsan Nowroozi; Samaneh Ghelichkhani; Imran Haider; Ali Dehghantanha A Survey on Transferability of Adversarial Examples across Deep Neural Networks. (99%)Jindong Gu; Xiaojun Jia; Jorge Pau de; Wenqain Yu; Xinwei Liu; Avery Ma; Yuan Xun; Anjun Hu; Ashkan Khakzar; Zhijiang Li; Xiaochun Cao; Philip Torr Defending Against Transfer Attacks From Public Models. (99%)Chawin Sitawarin; Jaewon Chang; David Huang; Wesson Altoyan; David Wagner Uncertainty-weighted Loss Functions for Improved Adversarial Attacks on Semantic Segmentation. (93%)Kira Maag; Asja Fischer Detection Defenses: An Empty Promise against Adversarial Patch Attacks on Optical Flow. (93%)Erik Scheurer; Jenny Schmalfuss; Alexander Lis; Andrés Bruhn CBD: A Certified Backdoor Detector Based on Local Dominant Probability. (76%)Zhen Xiang; Zidi Xiong; Bo Li SoK: Pitfalls in Evaluating Black-Box Attacks. (76%)Fnu Suya; Anshuman Suri; Tingwei Zhang; Jingtao Hong; Yuan Tian; David Evans Instability of computer vision models is a necessary result of the task itself. (26%)Oliver Turnbull; George Cevora PAC-tuning:Fine-tuning Pretrained Language Models with PAC-driven Perturbed Gradient Descent. (1%)Guangliang Liu; Zhiyu Xue; Xitong Zhang; Kristen Marie Johnson; Rongrong Wang A minimax optimal control approach for robust neural ODEs. (1%)Cristina Cipriani; Alessandro Scagliotti; Tobias Wöhrer 2023-10-25 Break it, Imitate it, Fix it: Robustness by Generating Human-Like Attacks. (93%)Aradhana Sinha; Ananth Balashankar; Ahmad Beirami; Thi Avrahami; Jilin Chen; Alex Beutel Trust, but Verify: Robust Image Segmentation using Deep Learning. (54%)Fahim Ahmed Zaman; Xiaodong Wu; Weiyu Xu; Milan Sonka; Raghuraman Mudumbai Dual Defense: Adversarial, Traceable, and Invisible Robust Watermarking against Face Swapping. (26%)Yunming Zhang; Dengpan Ye; Caiyun Xie; Long Tang; Chuanxi Chen; Ziyi Liu; Jiacheng Deng On the Proactive Generation of Unsafe Images From Text-To-Image Models Using Benign Prompts. (22%)Yixin Wu; Ning Yu; Michael Backes; Yun Shen; Yang Zhang Wide Flat Minimum Watermarking for Robust Ownership Verification of GANs. (12%)Jianwei Fei; Zhihua Xia; Benedetta Tondi; Mauro Barni Multi-scale Diffusion Denoised Smoothing. (1%)Jongheon Jeong; Jinwoo Shin SparseDFF: Sparse-View Feature Distillation for One-Shot Dexterous Manipulation. (1%)Qianxu Wang; Haotong Zhang; Congyue Deng; Yang You; Hao Dong; Yixin Zhu; Leonidas Guibas 2023-10-24 Adversarial sample generation and training using geometric masks for accurate and resilient license plate character recognition. (99%)Bishal Shrestha; Griwan Khakurel; Kritika Simkhada; Badri Adhikari RAEDiff: Denoising Diffusion Probabilistic Models Based Reversible Adversarial Examples Self-Generation and Self-Recovery. (92%)Fan Xing; Xiaoyi Zhou; Xuefeng Fan; Zhuo Tian; Yan Zhao Defense Against Model Extraction Attacks on Recommender Systems. (92%)Sixiao Zhang; Hongzhi Yin; Hongxu Chen; Cheng Long Segue: Side-information Guided Generative Unlearnable Examples for Facial Privacy Protection in Real World. (89%)Zhiling Zhang; Jie Zhang; Kui Zhang; Wenbo Zhou; Weiming Zhang; Nenghai Yu Hierarchical Randomized Smoothing. (75%)Yan Scholten; Jan Schuchardt; Aleksandar Bojchevski; Stephan Günnemann Momentum Gradient-based Untargeted Attack on Hypergraph Neural Networks. (73%)Yang Chen; Stjepan Picek; Zhonglin Ye; Zhaoyang Wang; Haixing Zhao Corrupting Neuron Explanations of Deep Visual Features. (41%)Divyansh Srivastava; Tuomas Oikarinen; Tsui-Wei Weng Guiding LLM to Fool Itself: Automatically Manipulating Machine Reading Comprehension Shortcut Triggers. (10%)Mosh Levy; Shauli Ravfogel; Yoav Goldberg A Survey on Detection of LLMs-Generated Content. (1%)Xianjun Yang; Liangming Pan; Xuandong Zhao; Haifeng Chen; Linda Petzold; William Yang Wang; Wei Cheng White-box Compiler Fuzzing Empowered by Large Language Models. (1%)Chenyuan Yang; Yinlin Deng; Runyu Lu; Jiayi Yao; Jiawei Liu; Reyhaneh Jabbarvand; Lingming Zhang Enhancing Large Language Models for Secure Code Generation: A Dataset-driven Study on Vulnerability Mitigation. (1%)Jiexin Wang; Liuwen Cao; Xitong Luo; Zhiping Zhou; Jiayuan Xie; Adam Jatowt; Yi Cai 2023-10-23 Semantic-Aware Adversarial Training for Reliable Deep Hashing Retrieval. (99%)Xu Yuan; Zheng Zhang; Xunguang Wang; Lin Wu F$^2$AT: Feature-Focusing Adversarial Training via Disentanglement of Natural and Perturbed Patterns. (99%)Yaguan Qian; Chenyu Zhao; Zhaoquan Gu; Bin Wang; Shouling Ji; Wei Wang; Boyang Zhou; Pan Zhou AutoDAN: Automatic and Interpretable Adversarial Attacks on Large Language Models. (98%)Sicheng Zhu; Ruiyi Zhang; Bang An; Gang Wu; Joe Barrow; Zichao Wang; Furong Huang; Ani Nenkova; Tong Sun Fast Propagation is Better: Accelerating Single-Step Adversarial Training via Sampling Subnetworks. (98%)Xiaojun Jia; Jianshu Li; Jindong Gu; Yang Bai; Xiaochun Cao On the Detection of Image-Scaling Attacks in Machine Learning. (15%)Erwin Quiring; Andreas Müller; Konrad Rieck RoboDepth: Robust Out-of-Distribution Depth Estimation under Corruptions. (1%)Lingdong Kong; Shaoyuan Xie; Hanjiang Hu; Lai Xing Ng; Benoit R. Cottereau; Wei Tsang Ooi 2023-10-22 Diffusion-Based Adversarial Purification for Speaker Verification. (99%)Yibo Bai; Xiao-Lei Zhang CT-GAT: Cross-Task Generative Adversarial Attack based on Transferability. (99%)Minxuan Lv; Chengwei Dai; Kun Li; Wei Zhou; Songlin Hu Imperceptible CMOS camera dazzle for adversarial attacks on deep neural networks. (92%)Zvi Stein; Adrian Stern ADoPT: LiDAR Spoofing Attack Detection Based on Point-Level Temporal Consistency. (26%)Minkyoung Cho; Yulong Cao; Zixiang Zhou; Z. Morley Mao Attention-Enhancing Backdoor Attacks Against BERT-based Models. (13%)Weimin Lyu; Songzhu Zheng; Lu Pang; Haibin Ling; Chao Chen MoPe: Model Perturbation-based Privacy Attacks on Language Models. (9%)Marvin Li; Jason Wang; Jeffrey Wang; Seth Neel Reputation-Based Federated Learning Defense to Mitigate Threats in EEG Signal Classification. (1%)Zhibo Zhang; Pengfei Li; Ahmed Y. Al Hammadi; Fusen Guo; Ernesto Damiani; Chan Yeob Yeun 2023-10-21 Adversarial Image Generation by Spatial Transformation in Perceptual Colorspaces. (99%)Ayberk Aydin; Alptekin Temizel Training Image Derivatives: Increased Accuracy and Universal Robustness. (5%)Vsevolod I. Avrutskiy 2023-10-20 Beyond Hard Samples: Robust and Effective Grammatical Error Correction with Cycle Self-Augmenting. (99%)Zecheng Tang; Kaifeng Qi; Juntao Li; Min Zhang An LLM can Fool Itself: A Prompt-Based Adversarial Attack. (99%)Xilie Xu; Keyi Kong; Ning Liu; Lizhen Cui; Di Wang; Jingfeng Zhang; Mohan Kankanhalli Prompt-Specific Poisoning Attacks on Text-to-Image Generative Models. (61%)Shawn Shan; Wenxin Ding; Josephine Passananti; Haitao Zheng; Ben Y. Zhao The Hidden Adversarial Vulnerabilities of Medical Federated Learning. (45%)Erfan Darzi; Florian Dubost; Nanna. M. Sijtsema; Ooijen P. M. A van Adversarial Attacks on Fairness of Graph Neural Networks. (26%)Binchi Zhang; Yushun Dong; Chen Chen; Yada Zhu; Minnan Luo; Jundong Li FLTracer: Accurate Poisoning Attack Provenance in Federated Learning. (26%)Xinyu Zhang; Qingyu Liu; Zhongjie Ba; Yuan Hong; Tianhang Zheng; Feng Lin; Li Lu; Kui Ren Can We Trust the Similarity Measurement in Federated Learning? (15%)Zhilin Wang; Qin Hu; Xukai Zou Data-Free Knowledge Distillation Using Adversarially Perturbed OpenGL Shader Images. (4%)Logan Frank; Jim Davis VOICE-ZEUS: Impersonating Zoom's E2EE-Protected Static Media and Textual Communications via Simple Voice Manipulations. (4%)Mashari Alatawi; Nitesh Saxena 2023-10-19 Generating Robust Adversarial Examples against Online Social Networks (OSNs). (98%)Jun Liu; Jiantao Zhou; Haiwei Wu; Weiwei Sun; Jinyu Tian Recoverable Privacy-Preserving Image Classification through Noise-like Adversarial Examples. (98%)Jun Liu; Jiantao Zhou; Jinyu Tian; Weiwei Sun Learn from the Past: A Proxy based Adversarial Defense Framework to Boost Robustness. (98%)Yaohua Liu; Jiaxin Gao; Zhu Liu; Xianghao Jiao; Xin Fan; Risheng Liu OODRobustBench: benchmarking and analyzing adversarial robustness under distribution shift. (97%)Lin Li; Yifei Wang; Chawin Sitawarin; Michael Spratling Automatic Hallucination Assessment for Aligned Large Language Models via Transferable Adversarial Attacks. (97%)Xiaodong Yu; Hao Cheng; Xiaodong Liu; Dan Roth; Jianfeng Gao PatchCURE: Improving Certifiable Robustness, Model Utility, and Computation Efficiency of Adversarial Patch Defenses. (97%)Chong Xiang; Tong Wu; Sihui Dai; Jonathan Petit; Suman Jana; Prateek Mittal Prompt Injection Attacks and Defenses in LLM-Integrated Applications. (47%)Yupei Liu; Yuqi Jia; Runpeng Geng; Jinyuan Jia; Neil Zhenqiang Gong Attack Prompt Generation for Red Teaming and Defending Large Language Models. (15%)Boyi Deng; Wenjie Wang; Fuli Feng; Yang Deng; Qifan Wang; Xiangnan He SecurityNet: Assessing Machine Learning Vulnerabilities on Public Models. (5%)Boyang Zhang; Zheng Li; Ziqing Yang; Xinlei He; Michael Backes; Mario Fritz; Yang Zhang To grok or not to grok: Disentangling generalization and memorization on corrupted algorithmic datasets. (1%)Darshil Doshi; Aritra Das; Tianyu He; Andrey Gromov Detecting Shared Data Manipulation in Distributed Optimization Algorithms. (1%)Mohannad Alkhraijah; Rachel Harris; Samuel Litchfield; David Huggins; Daniel K. Molzahn Towards Robust Pruning: An Adaptive Knowledge-Retention Pruning Strategy for Language Models. (1%)Jianwei Li; Qi Lei; Wei Cheng; Dongkuan Xu 2023-10-18 Exploring Decision-based Black-box Attacks on Face Forgery Detection. (99%)Zhaoyu Chen; Bo Li; Kaixun Jiang; Shuang Wu; Shouhong Ding; Wenqiang Zhang Segment Anything Meets Universal Adversarial Perturbation. (99%)Dongshen Han; Sheng Zheng; Chaoning Zhang IRAD: Implicit Representation-driven Image Resampling against Adversarial Attacks. (99%)Yue Cao; Tianlin Li; Xiaofeng Cao; Ivor Tsang; Yang Liu; Qing Guo Revisiting Transferable Adversarial Image Examples: Attack Categorization, Evaluation Guidelines, and New Insights. (99%)Zhengyu Zhao; Hanwei Zhang; Renjue Li; Ronan Sicre; Laurent Amsaleg; Michael Backes; Qi Li; Chao Shen Tailoring Adversarial Attacks on Deep Neural Networks for Targeted Class Manipulation Using DeepFool Algorithm. (99%)S. M. Fazle Rabby Labib; Joyanta Jyoti Mondal; Meem Arafat Manab Malicious Agent Detection for Robust Multi-Agent Collaborative Perception. (87%)Yangheng Zhao; Zhen Xiang; Sheng Yin; Xianghe Pang; Siheng Chen; Yanfeng Wang Black-Box Training Data Identification in GANs via Detector Networks. (82%)Lukman Olagoke; Salil Vadhan; Seth Neel Adversarial Training for Physics-Informed Neural Networks. (81%)Yao Li; Shengzhu Shi; Zhichang Guo; Boying Wu REVAMP: Automated Simulations of Adversarial Attacks on Arbitrary Objects in Realistic Scenes. (80%)Matthew Hull; Zijie J. Wang; Duen Horng Chau Quantifying Privacy Risks of Prompts in Visual Prompt Learning. (76%)Yixin Wu; Rui Wen; Michael Backes; Pascal Berrang; Mathias Humbert; Yun Shen; Yang Zhang To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Unsafe Images ... For Now. (47%)Yimeng Zhang; Jinghan Jia; Xin Chen; Aochuan Chen; Yihua Zhang; Jiancheng Liu; Ke Ding; Sijia Liu CAT: Closed-loop Adversarial Training for Safe End-to-End Driving. (2%)Linrui Zhang; Zhenghao Peng; Quanyi Li; Bolei Zhou PrivInfer: Privacy-Preserving Inference for Black-box Large Language Model. (1%)Meng Tong; Kejiang Chen; Yuang Qi; Jie Zhang; Weiming Zhang; Nenghai Yu 2023-10-17 The Efficacy of Transformer-based Adversarial Attacks in Security Domains. (99%)Kunyang Li; Kyle Domico; Jean-Charles Noirot Ferrand; Patrick McDaniel Adversarial Robustness Unhardening via Backdoor Attacks in Federated Learning. (93%)Taejin Kim; Jiarui Li; Shubhranshu Singh; Nikhil Madaan; Carlee Joe-Wong WaveAttack: Asymmetric Frequency Obfuscation-based Backdoor Attacks Against Deep Neural Networks. (15%)Jun Xia; Zhihao Yue; Yingbo Zhou; Zhiwei Ling; Xian Wei; Mingsong Chen Generalizability of CNN Architectures for Face Morph Presentation Attack. (1%)Sherko R. HmaSalah; Aras Asaad 2023-10-16 Survey of Vulnerabilities in Large Language Models Revealed by Adversarial Attacks. (98%)Erfan Shayegani; Md Abdullah Al Mamun; Yu Fu; Pedram Zaree; Yue Dong; Nael Abu-Ghazaleh Regularization properties of adversarially-trained linear regression. (92%)Antônio H. Ribeiro; Dave Zachariah; Francis Bach; Thomas B. Schön Fast Adversarial Label-Flipping Attack on Tabular Data. (84%)Xinglong Chang; Gillian Dobbie; Jörg Wicker A Non-monotonic Smooth Activation Function. (83%)Koushik Biswas; Meghana Karri; Ulaş Bağcı Quantifying Assistive Robustness Via the Natural-Adversarial Frontier. (68%)Jerry Zhi-Yang He; Zackory Erickson; Daniel S. Brown; Anca D. Dragan A Comprehensive Study of Privacy Risks in Curriculum Learning. (67%)Joann Qiongna Chen; Xinlei He; Zheng Li; Yang Zhang; Zhou Li DANAA: Towards transferable attacks with double adversarial neuron attribution. (26%)Zhibo Jin; Zhiyu Zhu; Xinyi Wang; Jiayu Zhang; Jun Shen; Huaming Chen Demystifying Poisoning Backdoor Attacks from a Statistical Perspective. (9%)Ganghua Wang; Xun Xian; Jayanth Srinivasa; Ashish Kundu; Xuan Bi; Mingyi Hong; Jie Ding Prompt Packer: Deceiving LLMs through Compositional Instruction with Hidden Attacks. (4%)Shuyu Jiang; Xingshu Chen; Rui Tang Robust Multi-Agent Reinforcement Learning via Adversarial Regularization: Theoretical Foundation and Stable Algorithms. (3%)Alexander Bukharin; Yan Li; Yue Yu; Qingru Zhang; Zhehui Chen; Simiao Zuo; Chao Zhang; Songan Zhang; Tuo Zhao Passive Inference Attacks on Split Learning via Adversarial Regularization. (3%)Xiaochen Zhu; Xinjian Luo; Yuncheng Wu; Yangfan Jiang; Xiaokui Xiao; Beng Chin Ooi On the Transferability of Learning Models for Semantic Segmentation for Remote Sensing Data. (2%)Rongjun Qin; Guixiang Zhang; Yang Tang Orthogonal Uncertainty Representation of Data Manifold for Robust Long-Tailed Learning. (1%)Yanbiao Ma; Licheng Jiao; Fang Liu; Shuyuan Yang; Xu Liu; Lingling Li Will the Prince Get True Love's Kiss? On the Model Sensitivity to Gender Perturbation over Fairytale Texts. (1%)Christina Chance; Da Yin; Dakuo Wang; Kai-Wei Chang 2023-10-15 Towards Deep Learning Models Resistant to Transfer-based Adversarial Attacks via Data-centric Robust Learning. (99%)Yulong Yang; Chenhao Lin; Xiang Ji; Qiwei Tian; Qian Li; Hongshan Yang; Zhibo Wang; Chao Shen SCME: A Self-Contrastive Method for Data-free and Query-Limited Model Extraction Attack. (99%)Renyang Liu; Jinhong Zhang; Kwok-Yan Lam; Jun Zhao; Wei Zhou AFLOW: Developing Adversarial Examples under Extremely Noise-limited Settings. (99%)Renyang Liu; Jinhong Zhang; Haoran Li; Jin Zhang; Yuanyu Wang; Wei Zhou Black-box Targeted Adversarial Attack on Segment Anything (SAM). (99%)Sheng Zheng; Chaoning Zhang; Xinhong Hao Evading Detection Actively: Toward Anti-Forensics against Forgery Localization. (97%)Long Zhuo; Shenghai Luo; Shunquan Tan; Han Chen; Bin Li; Jiwu Huang Explore the Effect of Data Selection on Poison Efficiency in Backdoor Attacks. (61%)Ziqiang Li; Pengfei Xia; Hong Sun; Yueqi Zeng; Wei Zhang; Bin Li Ring-A-Bell! How Reliable are Concept Removal Methods for Diffusion Models? (9%)Yu-Lin Tsai; Chia-Yi Hsu; Chulin Xie; Chih-Hsun Lin; Jia-You Chen; Bo Li; Pin-Yu Chen; Chia-Mu Yu; Chun-Ying Huang VFLAIR: A Research Library and Benchmark for Vertical Federated Learning. (3%)Tianyuan Zou; Zixuan Gu; Yu He; Hideaki Takahashi; Yang Liu; Ya-Qin Zhang 2023-10-14 BufferSearch: Generating Black-Box Adversarial Texts With Lower Queries. (98%)Wenjie Lv; Zhen Wang; Yitao Zheng; Zhehua Zhong; Qi Xuan; Tianyi Chen 2023-10-13 Is Certifying $\ell_p$ Robustness Still Worthwhile? (99%)Ravi Mangal; Klas Leino; Zifan Wang; Kai Hu; Weicheng Yu; Corina Pasareanu; Anupam Datta; Matt Fredrikson User Inference Attacks on Large Language Models. (41%)Nikhil Kandpal; Krishna Pillutla; Alina Oprea; Peter Kairouz; Christopher A. Choquette-Choo; Zheng Xu On the Over-Memorization During Natural, Robust and Catastrophic Overfitting. (1%)Runqi Lin; Chaojian Yu; Bo Han; Tongliang Liu 2023-10-12 Samples on Thin Ice: Re-Evaluating Adversarial Pruning of Neural Networks. (99%)Giorgio Piras; Maura Pintor; Ambra Demontis; Battista Biggio Concealed Electronic Countermeasures of Radar Signal with Adversarial Examples. (93%)Ruinan Ma; Canjie Zhu; Mingfeng Lu; Yunjie Li; Yu-an Tan; Ruibin Zhang; Ran Tao Attacks Meet Interpretability (AmI) Evaluation and Findings. (92%)Qian Ma; Ziping Ye; Shagufta Mehnaz Improving Fast Minimum-Norm Attacks with Hyperparameter Optimization. (68%)Giuseppe Floris; Raffaele Mura; Luca Scionis; Giorgio Piras; Maura Pintor; Ambra Demontis; Battista Biggio Fed-Safe: Securing Federated Learning in Healthcare Against Adversarial Attacks. (64%)Erfan Darzi; Nanna M. Sijtsema; Ooijen P. M. A van Provably Robust Cost-Sensitive Learning via Randomized Smoothing. (45%)Yuan Xin; Michael Backes; Xiao Zhang Bucks for Buckets (B4B): Active Defenses Against Stealing Encoders. (31%)Jan Dubiński; Stanisław Pawlak; Franziska Boenisch; Tomasz Trzciński; Adam Dziedzic Sentinel: An Aggregation Function to Secure Decentralized Federated Learning. (11%)Chao Feng; Alberto Huertas Celdran; Janosch Baltensperger; Enrique Tomas Matınez Bertran; Gerome Bovet; Burkhard Stiller Investigating the Robustness and Properties of Detection Transformers (DETR) Toward Difficult Images. (9%)Zhao Ning Zou; Yuhang Zhang; Robert Wijaya Defending Our Privacy With Backdoors. (9%)Dominik Hintersdorf; Lukas Struppek; Daniel Neider; Kristian Kersting Polynomial Time Cryptanalytic Extraction of Neural Network Models. (3%)Adi Shamir; Isaac Canales-Martinez; Anna Hambitzer; Jorge Chavez-Saab; Francisco Rodrigez-Henriquez; Nitin Satpute SEE-OoD: Supervised Exploration For Enhanced Out-of-Distribution Detection. (1%)Xiaoyang Song; Wenbo Sun; Maher Nouiehed; Raed Al Kontar; Judy Jin XAI Benchmark for Visual Explanation. (1%)Yifei Zhang; Siyi Gu; James Song; Bo Pan; Liang Zhao Jailbreaking Black Box Large Language Models in Twenty Queries. (1%)Patrick Chao; Alexander Robey; Edgar Dobriban; Hamed Hassani; George J. Pappas; Eric Wong Voyager: MTD-Based Aggregation Protocol for Mitigating Poisoning Attacks on DFL. (1%)Chao Feng; Alberto Huertas Celdran; Michael Vuong; Gerome Bovet; Burkhard Stiller 2023-10-11 Boosting Black-box Attack to Deep Neural Networks with Conditional Diffusion Models. (99%)Renyang Liu; Wei Zhou; Tianwei Zhang; Kangjie Chen; Jun Zhao; Kwok-Yan Lam Promoting Robustness of Randomized Smoothing: Two Cost-Effective Approaches. (89%)Linbo Liu; Trong Nghia Hoang; Lam M. Nguyen; Tsui-Wei Weng An Adversarial Example for Direct Logit Attribution: Memory Management in gelu-4l. (13%)James Dao; Yeu-Tong Lao; Can Rager; Jett Janiak Prompt Backdoors in Visual Prompt Learning. (11%)Hai Huang; Zhengyu Zhao; Michael Backes; Yun Shen; Yang Zhang Why Train More? Effective and Efficient Membership Inference via Memorization. (10%)Jihye Choi; Shruti Tople; Varun Chandrasekaran; Somesh Jha Towards Causal Deep Learning for Vulnerability Detection. (4%)Md Mahbubur Rahman; Ira Ceka; Chengzhi Mao; Saikat Chakraborty; Baishakhi Ray; Wei Le Deep Reinforcement Learning for Autonomous Cyber Operations: A Survey. (3%)Gregory Palmer; Chris Parry; Daniel J. B. Harrold; Chris Willis 2023-10-10 A Geometrical Approach to Evaluate the Adversarial Robustness of Deep Neural Networks. (99%)Yang Wang; Bo Dong; Ke Xu; Haiyin Piao; Yufei Ding; Baocai Yin; Xin Yang My Brother Helps Me: Node Injection Based Adversarial Attack on Social Bot Detection. (98%)Lanjun Wang; Xinran Qiao; Yanwei Xie; Weizhi Nie; Yongdong Zhang; Anan Liu Adversarial Robustness in Graph Neural Networks: A Hamiltonian Approach. (83%)Kai Zhao; Qiyu Kang; Yang Song; Rui She; Sijie Wang; Wee Peng Tay Adversarial optimization leads to over-optimistic security-constrained dispatch, but sampling can help. (76%)Charles Dawson; Chuchu Fan No Privacy Left Outside: On the (In-)Security of TEE-Shielded DNN Partition for On-Device ML. (62%)Ziqi Zhang; Chen Gong; Yifeng Cai; Yuanyuan Yuan; Bingyan Liu; Ding Li; Yao Guo; Xiangqun Chen Comparing the Robustness of Modern No-Reference Image- and Video-Quality Metrics to Adversarial Attacks. (47%)Anastasia Antsiferova; Khaled Abud; Aleksandr Gushchin; Ekaterina Shumitskaya; Sergey Lavrushkin; Dmitriy Vatolin GraphCloak: Safeguarding Task-specific Knowledge within Graph-structured Data from Unauthorized Exploitation. (22%)Yixin Liu; Chenrui Fan; Xun Chen; Pan Zhou; Lichao Sun Latent Diffusion Counterfactual Explanations. (5%)Karim Farid; Simon Schrodi; Max Argus; Thomas Brox FTFT: efficient and robust Fine-Tuning by transFerring Training dynamics. (2%)Yupei Du; Albert Gatt; Dong Nguyen Investigating the Adversarial Robustness of Density Estimation Using the Probability Flow ODE. (2%)Marius Arvinte; Cory Cornelius; Jason Martin; Nageen Himayat Jailbreak and Guard Aligned Language Models with Only Few In-Context Demonstrations. (1%)Zeming Wei; Yifei Wang; Yisen Wang 2023-10-09 PAC-Bayesian Spectrally-Normalized Bounds for Adversarially Robust Generalization. (92%)Jiancong Xiao; Ruoyu Sun; Zhi- Quan Luo Domain Watermark: Effective and Harmless Dataset Copyright Protection is Closed at Hand. (22%)Junfeng Guo; Yiming Li; Lixu Wang; Shu-Tao Xia; Heng Huang; Cong Liu; Bo Li Theoretical Analysis of Robust Overfitting for Wide DNNs: An NTK Approach. (5%)Shaopeng Fu; Di Wang Exploring adversarial attacks in federated learning for medical imaging. (2%)Erfan Darzi; Florian Dubost; N. M. Sijtsema; Ooijen P. M. A van 2023-10-08 An Initial Investigation of Neural Replay Simulator for Over-the-Air Adversarial Perturbations to Automatic Speaker Verification. (99%)Jiaqi Li; Li Wang; Liumeng Xue; Lei Wang; Zhizheng Wu AdvSV: An Over-the-Air Adversarial Attack Dataset for Speaker Verification. (96%)Li Wang; Jiaqi Li; Yuhao Luo; Jiahao Zheng; Lei Wang; Hao Li; Ke Xu; Chengfang Fang; Jie Shi; Zhizheng Wu BRAINTEASER: Lateral Thinking Puzzles for Large Language Models. (26%)Yifan Jiang; Filip Ilievski; Kaixin Ma; Zhivar Sourati 2023-10-07 Improving Adversarial Attacks on Latent Diffusion Model. (99%)Boyang Zheng; Chumeng Liang; Xiaoyu Wu; Yan Liu IPMix: Label-Preserving Data Augmentation Method for Training Robust Classifiers. (76%)Zhenglin Huang; Xiaoan Bao; Na Zhang; Qingqi Zhang; Xiaomei Tu; Biao Wu; Xi Yang 2023-10-06 VLATTACK: Multimodal Adversarial Attacks on Vision-Language Tasks via Pre-trained Models. (98%)Ziyi Yin; Muchao Ye; Tianrong Zhang; Tianyu Du; Jinguo Zhu; Han Liu; Jinghui Chen; Ting Wang; Fenglong Ma Kick Bad Guys Out! Zero-Knowledge-Proof-Based Anomaly Detection in Federated Learning. (84%)Shanshan Han; Wenxuan Wu; Baturalp Buyukates; Weizhao Jin; Qifan Zhang; Yuhang Yao; Salman Avestimehr; Chaoyang He 2023-10-05 OMG-ATTACK: Self-Supervised On-Manifold Generation of Transferable Evasion Attacks. (99%)Ofir Bar Tal; Adi Haviv; Amit H. Bermano Untargeted White-box Adversarial Attack with Heuristic Defence Methods in Real-time Deep Learning based Network Intrusion Detection System. (99%)Khushnaseeb Roshan; Aasim Zafar; Sheikh Burhan Ul Haque Enhancing Robust Representation in Adversarial Training: Alignment and Exclusion Criteria. (99%)Nuoyan Zhou; Nannan Wang; Decheng Liu; Dawei Zhou; Xinbo Gao An Integrated Algorithm for Robust and Imperceptible Audio Adversarial Examples. (98%)Armin Ettenhofer; Jan-Philipp Schulze; Karla Pizzi Adversarial Machine Learning for Social Good: Reframing the Adversary as an Ally. (98%)Shawqi Al-Maliki; Adnan Qayyum; Hassan Ali; Mohamed Abdallah; Junaid Qadir; Dinh Thai Hoang; Dusit Niyato; Ala Al-Fuqaha SmoothLLM: Defending Large Language Models Against Jailbreaking Attacks. (92%)Alexander Robey; Eric Wong; Hamed Hassani; George J. Pappas Targeted Adversarial Attacks on Generalizable Neural Radiance Fields. (56%)Andras Horvath; Csaba M. Jozsa Certification of Deep Learning Models for Medical Image Segmentation. (15%)Othmane Laousy; Alexandre Araujo; Guillaume Chassagnon; Nikos Paragios; Marie-Pierre Revel; Maria Vakalopoulou Certifiably Robust Graph Contrastive Learning. (5%)Minhua Lin; Teng Xiao; Enyan Dai; Xiang Zhang; Suhang Wang Towards Robust and Generalizable Training: An Empirical Study of Noisy Slot Filling for Input Perturbations. (2%)Jiachi Liu; Liwen Wang; Guanting Dong; Xiaoshuai Song; Zechen Wang; Zhengyang Wang; Shanglin Lei; Jinzheng Zhao; Keqing He; Bo Xiao; Weiran Xu 2023-10-04 Optimizing Key-Selection for Face-based One-Time Biometrics via Morphing. (98%)Daile Osorio-Roig; Mahdi Ghafourian; Christian Rathgeb; Ruben Vera-Rodriguez; Christoph Busch; Julian Fierrez Misusing Tools in Large Language Models With Visual Adversarial Examples. (97%)Xiaohan Fu; Zihan Wang; Shuheng Li; Rajesh K. Gupta; Niloofar Mireshghallah; Taylor Berg-Kirkpatrick; Earlence Fernandes Burning the Adversarial Bridges: Robust Windows Malware Detection Against Binary-level Mutations. (82%)Ahmed Abusnaina; Yizhen Wang; Sunpreet Arora; Ke Wang; Mihai Christodorescu; David Mohaisen Raze to the Ground: Query-Efficient Adversarial HTML Attacks on Machine-Learning Phishing Webpage Detectors. (81%)Biagio Montaruli; Luca Demetrio; Maura Pintor; Luca Compagna; Davide Balzarotti; Battista Biggio Shielding the Unseen: Privacy Protection through Poisoning NeRF with Spatial Deformation. (10%)Yihan Wu; Brandon Y. Feng; Heng Huang 2023-10-03 Splitting the Difference on Adversarial Training. (99%)Matan Levi; Aryeh Kontorovich DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training. (97%)Aochuan Chen; Yimeng Zhang; Jinghan Jia; James Diffenderfer; Jiancheng Liu; Konstantinos Parasyris; Yihua Zhang; Zheng Zhang; Bhavya Kailkhura; Sijia Liu SlowFormer: Universal Adversarial Patch for Attack on Compute and Energy Efficiency of Inference Efficient Vision Transformers. (86%)KL Navaneet; Soroush Abbasi Koohpayegani; Essam Sleiman; Hamed Pirsiavash Towards Stable Backdoor Purification through Feature Shift Tuning. (83%)Rui Min; Zeyu Qin; Li Shen; Minhao Cheng Jailbreaker in Jail: Moving Target Defense for Large Language Models. (73%)Bocheng Chen; Advait Paliwal; Qiben Yan AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models. (56%)Xiaogeng Liu; Nan Xu; Muhao Chen; Chaowei Xiao Beyond Labeling Oracles: What does it mean to steal ML models? (47%)Avital Shafran; Ilia Shumailov; Murat A. Erdogdu; Nicolas Papernot Exploring Model Learning Heterogeneity for Boosting Ensemble Robustness. (13%)Yanzhao Wu; Ka-Ho Chow; Wenqi Wei; Ling Liu FLEDGE: Ledger-based Federated Learning Resilient to Inference and Backdoor Attacks. (11%)Jorge Castillo; Phillip Rieger; Hossein Fereidooni; Qian Chen; Ahmad Sadeghi AutoLoRa: A Parameter-Free Automated Robust Fine-Tuning Framework. (3%)Xilie Xu; Jingfeng Zhang; Mohan Kankanhalli 2023-10-02 Fooling the Textual Fooler via Randomizing Latent Representations. (99%)Duy C. Hoang; Quang H. Nguyen; Saurav Manchanda; MinLong Peng; Kok-Seng Wong; Khoa D. Doan Adversarial Client Detection via Non-parametric Subspace Monitoring in the Internet of Federated Things. (92%)Xianjian Xie; Xiaochen Xian; Dan Li; Andi Wang LoFT: Local Proxy Fine-tuning For Improving Transferability Of Adversarial Attacks Against Large Language Model. (87%)Muhammad Ahmed Shah; Roshan Sharma; Hira Dhamyal; Raphael Olivier; Ankit Shah; Joseph Konan; Dareen Alharthi; Hazim T Bukhari; Massa Baali; Soham Deshmukh; Michael Kuhlmann; Bhiksha Raj; Rita Singh LLM Lies: Hallucinations are not Bugs, but Features as Adversarial Examples. (87%)Jia-Yu Yao; Kun-Peng Ning; Zhen-Hui Liu; Mu-Nan Ning; Li Yuan Gotcha! This Model Uses My Code! Evaluating Membership Leakage Risks in Code Models. (13%)Zhou Yang; Zhipeng Zhao; Chenyu Wang; Jieke Shi; Dongsum Kim; Donggyun Han; David Lo Toward effective protection against diffusion based mimicry through score distillation. (3%)Haotian Xue; Chumeng Liang; Xiaoyu Wu; Yongxin Chen Fool Your (Vision and) Language Model With Embarrassingly Simple Permutations. (1%)Yongshuo Zong; Tingyang Yu; Bingchen Zhao; Ruchika Chavhan; Timothy Hospedales 2023-10-01 A Survey of Robustness and Safety of 2D and 3D Deep Learning Models Against Adversarial Attacks. (99%)Yanjie Li; Bin Xie; Songtao Guo; Yuanyuan Yang; Bin Xiao Counterfactual Image Generation for adversarially robust and interpretable Classifiers. (96%)Rafael Bischof; Florian Scheidegger; Michael A. Kraus; A. Cristiano I. Malossi On the Onset of Robust Overfitting in Adversarial Training. (64%)Chaojian Yu; Xiaolong Shi; Jun Yu; Bo Han; Tongliang Liu Understanding Adversarial Transferability in Federated Learning. (64%)Yijiang Li; Ying Gao; Haohan Wang GhostEncoder: Stealthy Backdoor Attacks with Dynamic Triggers to Pre-trained Encoders in Self-supervised Learning. (61%)Qiannan Wang; Changchun Yin; Zhe Liu; Liming Fang; Run Wang; Chenhao Lin Fewer is More: Trojan Attacks on Parameter-Efficient Fine-Tuning. (9%)Lauren Hong; Ting Wang Can Pre-trained Networks Detect Familiar Out-of-Distribution Data? (1%)Atsuyuki Miyai; Qing Yu; Go Irie; Kiyoharu Aizawa How well does LLM generate security tests? (1%)Ying Daphne Zhang; Wenjia Daphne Song; Zhengjie Daphne Ji; Daphne Danfeng; Yao; Na Meng 2023-09-30 Understanding the Robustness of Randomized Feature Defense Against Query-Based Adversarial Attacks. (99%)Quang H. Nguyen; Yingjie Lao; Tung Pham; Kok-Seng Wong; Khoa D. Doan Human-Producible Adversarial Examples. (98%)David Khachaturov; Yue Gao; Ilia Shumailov; Robert Mullins; Ross Anderson; Kassem Fawaz Black-box Attacks on Image Activity Prediction and its Natural Language Explanations. (98%)Alina Elena Baia; Valentina Poggioni; Andrea Cavallaro Horizontal Class Backdoor to Deep Learning. (56%)Hua Ma; Shang Wang; Yansong Gao Refutation of Shapley Values for XAI -- Additional Evidence. (8%)Xuanxiang Huang; Joao Marques-Silva 2023-09-29 Robustness of AI-Image Detectors: Fundamental Limits and Practical Attacks. (99%)Mehrdad Saberi; Vinu Sankar Sadasivan; Keivan Rezaei; Aounon Kumar; Atoosa Chegini; Wenxiao Wang; Soheil Feizi Efficient Biologically Plausible Adversarial Training. (98%)Matilde Tristany Farinha; Thomas Ortner; Giorgia Dellaferrera; Benjamin Grewe; Angeliki Pantazi Can Sensitive Information Be Deleted From LLMs? Objectives for Defending Against Extraction Attacks. (96%)Vaidehi Patil; Peter Hase; Mohit Bansal On Continuity of Robust and Accurate Classifiers. (93%)Ramin Barati; Reza Safabakhsh; Mohammad Rahmati Adversarial Machine Learning in Latent Representations of Neural Networks. (93%)Milin Zhang; Mohammad Abdi; Francesco Restuccia Certified Robustness via Dynamic Margin Maximization and Improved Lipschitz Regularization. (92%)Mahyar Fazlyab; Taha Entesari; Aniket Roy; Rama Chellappa Toward Robust Recommendation via Real-time Vicinal Defense. (82%)Yichang Xu; Chenwang Wu; Defu Lian Adversarial Explainability: Utilizing Explainable Machine Learning in Bypassing IoT Botnet Detection Systems. (31%)Mohammed M. Alani; Atefeh Mashatan; Ali Miri Practical Membership Inference Attacks Against Large-Scale Multi-Modal Models: A Pilot Study. (13%)Myeongseob Ko; Ming Jin; Chenguang Wang; Ruoxi Jia Distributed Resilient Control of DC Microgrids Under Generally Unbounded FDI Attacks. (1%)Yichao Wang; Mohamadamin Rajabinezhad; Omar A. Beg; Shan Zuo Source Inference Attacks: Beyond Membership Inference Attacks in Federated Learning. (1%)Hongsheng Hu; Xuyun Zhang; Zoran Salcic; Lichao Sun; Kim-Kwang Raymond Choo; Gillian Dobbie 2023-09-28 Investigating Human-Identifiable Features Hidden in Adversarial Perturbations. (98%)Dennis Y. Menn; Tzu-hsun Feng; Sriram Vishwanath; Hung-yi Lee Parameter-Saving Adversarial Training: Reinforcing Multi-Perturbation Robustness via Hypernetworks. (98%)Huihui Gong; Minjing Dong; Siqi Ma; Seyit Camtepe; Surya Nepal; Chang Xu Towards Poisoning Fair Representations. (70%)Tianci Liu; Haoyu Wang; Feijie Wu; Hengtong Zhang; Pan Li; Lu Su; Jing Gao On the Trade-offs between Adversarial Robustness and Actionable Explanations. (68%)Satyapriya Krishna; Chirag Agarwal; Himabindu Lakkaraju The Lipschitz-Variance-Margin Tradeoff for Enhanced Randomized Smoothing. (56%)Blaise Delattre; Alexandre Araujo; Quentin Barthélemy; Alexandre Allauzen Post-Training Overfitting Mitigation in DNN Classifiers. (41%)Hang Wang; David J. Miller; George Kesidis Leveraging Optimization for Adaptive Attacks on Image Watermarks. (13%)Nils Lukas; Abdulrahman Diaa; Lucas Fenaux; Florian Kerschbaum Random and Safe Cache Architecture to Defeat Cache Timing Attacks. (9%)Guangyuan Hu; Ruby B. Lee Robust Offline Reinforcement Learning -- Certify the Confidence Interval. (4%)Jiarui Yao; Simon Shaolei Du A Primer on Bayesian Neural Networks: Review and Debates. (2%)Julyan Arbel; Konstantinos Pitas; Mariia Vladimirova; Vincent Fortuin 2023-09-27 Adversarial Examples Might be Avoidable: The Role of Data Concentration in Adversarial Robustness. (95%)Ambar Pal; Jeremias Sulam; René Vidal Defending Against Physical Adversarial Patch Attacks on Infrared Human Detection. (95%)Lukas Strack; Futa Waseda; Huy H. Nguyen; Yinqiang Zheng; Isao Echizen On the Computational Entanglement of Distant Features in Adversarial Machine Learning. (92%)YenLung Lai; Xingbo Dong; Zhe Jin Automatic Feature Fairness in Recommendation via Adversaries. (33%)Hengchang Hu; Yiming Cao; Zhankui He; Samson Tan; Min-Yen Kan Warfare:Breaking the Watermark Protection of AI-Generated Content. (12%)Guanlin Li; Yifei Chen; Jie Zhang; Jiwei Li; Shangwei Guo; Tianwei Zhang Generating Transferable Adversarial Simulation Scenarios for Self-Driving via Neural Rendering. (11%)Yasasa Abeysirigoonawardena; Kevin Xie; Chuhan Chen; Salar Hosseini; Ruiting Chen; Ruiqi Wang; Florian Shkurti Breaking On-Chip Communication Anonymity using Flow Correlation Attacks. (4%)Hansika Weerasena; Prabhat Mishra Genetic Algorithm-Based Dynamic Backdoor Attack on Federated Learning-Based Network Traffic Classification. (1%)Mahmoud Nazzal; Nura Aljaafari; Ahmed Sawalmeh; Abdallah Khreishah; Muhammad Anan; Abdulelah Algosaibi; Mohammed Alnaeem; Adel Aldalbahi; Abdulaziz Alhumam; Conrado P. Vizcarra; Shadan Alhamed 2023-09-26 Structure Invariant Transformation for better Adversarial Transferability. (99%)Xiaosen Wang; Zeliang Zhang; Jianping Zhang Privacy-preserving and Privacy-attacking Approaches for Speech and Audio -- A Survey. (16%)Yuchen Liu; Apu Kapadia; Donald Williamson Neural Stochastic Differential Equations for Robust and Explainable Analysis of Electromagnetic Unintended Radiated Emissions. (2%)Sumit Kumar Jha; Susmit Jha; Rickard Ewetz; Alvaro Velasquez Collaborative Watermarking for Adversarial Speech Synthesis. (1%)Lauri Aalto University, Finland Juvela; Xin National Institute of Informatics, Japan Wang 2023-09-25 DifAttack: Query-Efficient Black-Box Attack via Disentangled Feature Space. (99%)Liu Jun; Zhou Jiantao; Zeng Jiandian; Jinyu Tian Gray-box Adversarial Attack of Deep Reinforcement Learning-based Trading Agents. (98%)Foozhan Ataiefard; Hadi Hemmati SurrogatePrompt: Bypassing the Safety Filter of Text-To-Image Models via Substitution. (1%)Zhongjie Ba; Jieming Zhong; Jiachen Lei; Peng Cheng; Qinglong Wang; Zhan Qin; Zhibo Wang; Kui Ren 2023-09-24 Adversarial Attacks on Video Object Segmentation with Hard Region Discovery. (99%)Ping Li; Yu Zhang; Li Yuan; Jian Zhao; Xianghua Xu; Xiaoqin Zhang Vulnerabilities in Video Quality Assessment Models: The Challenge of Adversarial Attacks. (98%)Ao-Xiang Zhang; Yu Ran; Weixuan Tang; Yuan-Gen Wang On the Effectiveness of Adversarial Samples against Ensemble Learning-based Windows PE Malware Detectors. (86%)Trong-Nghia To; Danh Le Kim; Do Thi Thu Hien; Nghi Hoang Khoa; Hien Do Hoang; Phan The Duy; Van-Hau Pham Benchmarking Local Robustness of High-Accuracy Binary Neural Networks for Enhanced Traffic Sign Recognition. (80%)Andreea Postovan; Mădălina Eraşcu Projected Randomized Smoothing for Certified Adversarial Robustness. (76%)Samuel Pfrommer; Brendon G. Anderson; Somayeh Sojoudi Combining Two Adversarial Attacks Against Person Re-Identification Systems. (73%)Eduardo de O. Andrade; Igor Garcia Ballhausen Sampaio; Joris Guérin; José Viterbo Seeing Is Not Always Believing: Invisible Collision Attack and Defence on Pre-Trained Models. (2%)Minghang Deng; Zhong Zhang; Junming Shao 2023-09-23 Defending Pre-trained Language Models as Few-shot Learners against Backdoor Attacks. (61%)Zhaohan Xi; Tianyu Du; Changjiang Li; Ren Pang; Shouling Ji; Jinghui Chen; Fenglong Ma; Ting Wang Moving Target Defense based Secured Network Slicing System in the O-RAN Architecture. (1%)Mojdeh Karbalaee Motalleb; Chafika Benzaïd; Tarik Taleb; Vahid Shah-Mansouri Detecting and Mitigating System-Level Anomalies of Vision-Based Controllers. (1%)Aryaman Gupta; Kaustav Chakraborty; Somil Bansal 2023-09-22 RBFormer: Improve Adversarial Robustness of Transformer by Robust Bias. (99%)Hao Cheng; Jinhao Duan; Hui Li; Lyutianyang Zhang; Jiahang Cao; Ping Wang; Jize Zhang; Kaidi Xu; Renjing Xu Spatial-frequency channels, shape bias, and adversarial robustness. (69%)Ajay Subramanian; Elena Sizikova; Najib J. Majaj; Denis G. Pelli VIC-KD: Variance-Invariance-Covariance Knowledge Distillation to Make Keyword Spotting More Robust Against Adversarial Attacks. (69%)Heitor R. Guimarães; Arthur Pimentel; Anderson Avila; Tiago H. Falk Understanding Deep Gradient Leakage via Inversion Influence Functions. (15%)Haobo Zhang; Junyuan Hong; Yuyang Deng; Mehrdad Mahdavi; Jiayu Zhou Pixel-wise Smoothing for Certified Robustness against Camera Motion Perturbations. (10%)Hanjiang Hu; Zuxin Liu; Linyi Li; Jiacheng Zhu; Ding Zhao Privacy Assessment on Reconstructed Images: Are Existing Evaluation Metrics Faithful to Human Perception? (5%)Xiaoxiao Sun; Nidham Gazagnadou; Vivek Sharma; Lingjuan Lyu; Hongdong Li; Liang Zheng Expressive variational quantum circuits provide inherent privacy in federated learning. (1%)Niraj Kumar; Jamie Heredge; Changhao Li; Shaltiel Eloul; Shree Hari Sureshbabu; Marco Pistoia On Data Fabrication in Collaborative Vehicular Perception: Attacks and Countermeasures. (1%)Qingzhao Zhang; Shuowei Jin; Ruiyang Zhu; Jiachen Sun; Xumiao Zhang; Qi Alfred Chen; Z. Morley Mao 2023-09-21 Improving Machine Learning Robustness via Adversarial Training. (99%)Long Dang; Thushari Hapuarachchi; Kaiqi Xiong; Jing Lin Goal-Oriented Prompt Attack and Safety Evaluation for LLMs. (69%)Chengyuan Liu; Fubang Zhao; Lizhi Qing; Yangyang Kang; Changlong Sun; Kun Kuang; Fei Wu HANS, are you clever? Clever Hans Effect Analysis of Neural Systems. (45%)Leonardo Ranaldi; Fabio Massimo Zanzotto On the Relationship between Skill Neurons and Robustness in Prompt Tuning. (12%)Leon Ackermann; Xenia Ohmer DeepTheft: Stealing DNN Model Architectures through Power Side Channel. (1%)Yansong Gao; Huming Qiu; Zhi Zhang; Binghui Wang; Hua Ma; Alsharif Abuadbba; Minhui Xue; Anmin Fu; Surya Nepal 2023-09-20 How Robust is Google's Bard to Adversarial Image Attacks? (99%)Yinpeng Dong; Huanran Chen; Jiawei Chen; Zhengwei Fang; Xiao Yang; Yichi Zhang; Yu Tian; Hang Su; Jun Zhu PRAT: PRofiling Adversarial aTtacks. (99%)Rahul Ambati; Naveed Akhtar; Ajmal Mian; Yogesh Singh Rawat When to Trust AI: Advances and Challenges for Certification of Neural Networks. (64%)Marta Kwiatkowska; Xiyue Zhang AudioFool: Fast, Universal and synchronization-free Cross-Domain Attack on Speech Recognition. (54%)Mohamad Fakih; Rouwaida Kanj; Fadi Kurdahi; Mohammed E. Fouda Understanding Pose and Appearance Disentanglement in 3D Human Pose Estimation. (54%)Krishna Kanth Nakka; Mathieu Salzmann Fed-LSAE: Thwarting Poisoning Attacks against Federated Cyber Threat Detection System via Autoencoder-based Latent Space Inspection. (5%)Tran Duc Luong; Vuong Minh Tien; Nguyen Huu Quyen; Do Thi Thu Hien; Phan The Duy; Van-Hau Pham Compilation as a Defense: Enhancing DL Model Attack Robustness via Tensor Optimization. (2%)Stefan Trawicki; William Hackett; Lewis Birch; Neeraj Suri; Peter Garraghan 2023-09-19 Language Guided Adversarial Purification. (99%)Himanshu Singh; A V Subramanyam What Learned Representations and Influence Functions Can Tell Us About Adversarial Examples. (99%)Shakila Mahjabin Tonni; Mark Dras Adversarial Attacks Against Uncertainty Quantification. (99%)Emanuele Ledda; Daniele Angioni; Giorgio Piras; Giorgio Fumera; Battista Biggio; Fabio Roli Model Leeching: An Extraction Attack Targeting LLMs. (76%)Lewis Birch; William Hackett; Stefan Trawicki; Neeraj Suri; Peter Garraghan Information Leakage from Data Updates in Machine Learning Models. (16%)Tian Hui; Farhad Farokhi; Olga Ohrimenko Robin: A Novel Method to Produce Robust Interpreters for Deep Learning-Based Code Classifiers. (16%)Zhen Li; Ruqian Zhang; Deqing Zou; Ning Wang; Yating Li; Shouhuai Xu; Chen Chen; Hai Jin SPFL: A Self-purified Federated Learning Method Against Poisoning Attacks. (12%)Zizhen Liu; Weiyang He; Chip-Hong Chang; Jing Ye; Huawei Li; Xiaowei Li It's Simplex! Disaggregating Measures to Improve Certified Robustness. (11%)Andrew C. Cullen; Paul Montague; Shijie Liu; Sarah M. Erfani; Benjamin I. P. Rubinstein Nebula: Self-Attention for Dynamic Malware Analysis. (5%)Dmitrijs Trizna; Luca Demetrio; Battista Biggio; Fabio Roli Extreme Image Transformations Facilitate Robust Latent Object Representations. (1%)Girik Malik; Dakarai Crowder; Ennio Mingolla 2023-09-18 Stealthy Physical Masked Face Recognition Attack via Adversarial Style Optimization. (99%)Huihui Gong; Minjing Dong; Siqi Ma; Seyit Camtepe; Surya Nepal; Chang Xu Transferable Adversarial Attack on Image Tampering Localization. (99%)Yuqi Wang; Gang Cao; Zijie Lou; Haochen Zhu Efficient Low-Rank GNN Defense Against Structural Attacks. (96%)Abdullah Alchihabi; Qing En; Yuhong Guo Evaluating Adversarial Robustness with Expected Viable Performance. (45%)Ryan McCoppin; Colin Dawson; Sean M. Kennedy; Leslie M. Blaha Dual Student Networks for Data-Free Model Stealing. (26%)James Beetham; Navid Kardan; Ajmal Mian; Mubarak Shah Securing Fixed Neural Network Steganography. (5%)Zicong Luo; Sheng Li; Guobiao Li; Zhenxing Qian; Xinpeng Zhang GPTFUZZER: Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts. (4%)Jiahao Yu; Xingwei Lin; Zheng Yu; Xinyu Xing Spoofing attack augmentation: can differently-trained attack models improve generalisation? (3%)Wanying Ge; Xin Wang; Junichi Yamagishi; Massimiliano Todisco; Nicholas Evans Frame-to-Utterance Convergence: A Spectra-Temporal Approach for Unified Spoofing Detection. (1%)Awais Khan; Khalid Mahmood Malik; Shah Nawaz 2023-09-17 Reducing Adversarial Training Cost with Gradient Approximation. (99%)Huihui Gong; Shuo Yang; Siqi Ma; Seyit Camtepe; Surya Nepal; Chang Xu Defending Against Alignment-Breaking Attacks via Robustly Aligned LLM. (61%)Bochuan Cao; Yuanpu Cao; Lu Lin; Jinghui Chen 2023-09-16 Context-aware Adversarial Attack on Named Entity Recognition. (99%)Shuguang Chen; Leonardo Neves; Thamar Solorio Inverse classification with logistic and softmax classifiers: efficient optimization. (56%)Miguel Á. Carreira-Perpiñán; Suryabhan Singh Hada Robust Backdoor Attacks on Object Detection in Real World. (11%)Yaguan Qian; Boyuan Ji; Shuke He; Shenhui Huang; Xiang Ling; Bin Wang; Wei Wang Conditional Mutual Information Constrained Deep Learning for Classification. (5%)En-Hui Yang; Shayan Mohajer Hamidi; Linfeng Ye; Renhao Tan; Beverly Yang 2023-09-15 Adversarial Attacks on Tables with Entity Swap. (92%)Aneta Koleva; Martin Ringsquandl; Volker Tresp HINT: Healthy Influential-Noise based Training to Defend against Data Poisoning Attacks. (87%)Minh-Hao Van; Alycia N. Carey; Xintao Wu Distributionally Robust Post-hoc Classifiers under Prior Shifts. (1%)Jiaheng Wei; Harikrishna Narasimhan; Ehsan Amid; Wen-Sheng Chu; Yang Liu; Abhishek Kumar A Duty to Forget, a Right to be Assured? Exposing Vulnerabilities in Machine Unlearning Services. (1%)Hongsheng Hu; Shuo Wang; Jiamin Chang; Haonan Zhong; Ruoxi Sun; Shuang Hao; Haojin Zhu; Minhui Xue 2023-09-14 Unleashing the Adversarial Facet of Software Debloating. (98%)Do-Men Su; Mohannad Alhanahnah SLMIA-SR: Speaker-Level Membership Inference Attacks against Speaker Recognition Systems. (76%)Guangke Chen; Yedi Zhang; Fu Song What Matters to Enhance Traffic Rule Compliance of Imitation Learning for Automated Driving. (50%)Hongkuan Zhou; Aifen Sui; Wei Cao; Zhenshan Bing BAGEL: Backdoor Attacks against Federated Contrastive Learning. (16%)Yao Huang; Kongyang Chen; Jiannong Cao; Jiaxing Shen; Shaowei Wang; Yun Peng; Weilong Peng; Kechao Cai Physical Invisible Backdoor Based on Camera Imaging. (2%)Yusheng Guo; Nan Zhong; Zhenxing Qian; Xinpeng Zhang M3Dsynth: A dataset of medical 3D images with AI-generated local manipulations. (1%)Giada Zingarini; Davide Cozzolino; Riccardo Corvi; Giovanni Poggi; Luisa Verdoliva 2023-09-13 Semantic Adversarial Attacks via Diffusion Models. (99%)Chenan Wang; Jinhao Duan; Chaowei Xiao; Edward Kim; Matthew Stamm; Kaidi Xu Hardening RGB-D Object Recognition Systems against Adversarial Patch Attacks. (99%)Yang Zheng; Luca Demetrio; Antonio Emanuele Cinà; Xiaoyi Feng; Zhaoqiang Xia; Xiaoyue Jiang; Ambra Demontis; Battista Biggio; Fabio Roli Mitigating Adversarial Attacks in Federated Learning with Trusted Execution Environments. (99%)Simon Queyrut; Valerio Schiavoni; Pascal Felber PhantomSound: Black-Box, Query-Efficient Audio Adversarial Attack via Split-Second Phoneme Injection. (99%)Hanqing Guo; Guangjing Wang; Yuanda Wang; Bocheng Chen; Qiben Yan; Li Xiao APICom: Automatic API Completion via Prompt Learning and Adversarial Training-based Data Augmentation. (92%)Yafeng Gu; Yiheng Shen; Xiang Chen; Shaoyu Yang; Yiling Huang; Zhixiang Cao RAIN: Your Language Models Can Align Themselves without Finetuning. (83%)Yuhui Li; Fangyun Wei; Jinjing Zhao; Chao Zhang; Hongyang Zhang Differentiable JPEG: The Devil is in the Details. (70%)Christoph Reich; Biplob Debnath; Deep Patel; Srimat Chakradhar Deep Nonparametric Convexified Filtering for Computational Photography, Image Synthesis and Adversarial Defense. (41%)Jianqiao Wangni MASTERKEY: Practical Backdoor Attack Against Speaker Verification Systems. (38%)Hanqing Guo; Xun Chen; Junfeng Guo; Li Xiao; Qiben Yan Client-side Gradient Inversion Against Federated Learning from Poisoning. (22%)Jiaheng Wei; Yanjun Zhang; Leo Yu Zhang; Chao Chen; Shirui Pan; Kok-Leong Ong; Jun Zhang; Yang Xiang Safe Reinforcement Learning with Dual Robustness. (1%)Zeyang Li; Chuxiong Hu; Yunan Wang; Yujie Yang; Shengbo Eben Li 2023-09-12 Using Reed-Muller Codes for Classification with Rejection and Recovery. (99%)Daniel University of Birmingham Fentham; David University of Oxford Parker; Mark University of Birmingham Ryan Certified Robust Models with Slack Control and Large Lipschitz Constants. (98%)Max Losch; David Stutz; Bernt Schiele; Mario Fritz Exploring Non-additive Randomness on ViT against Query-Based Black-Box Attacks. (98%)Jindong Gu; Fangyun Wei; Philip Torr; Han Hu Backdoor Attacks and Countermeasures in Natural Language Processing Models: A Comprehensive Security Review. (61%)Pengzhou Cheng; Zongru Wu; Wei Du; Gongshen Liu CToMP: A Cycle-task-oriented Memory Protection Scheme for Unmanned Systems. (8%)Chengyan Ma; Ning Xi; Di Lu; Yebo Feng; Jianfeng Ma Language Models as Black-Box Optimizers for Vision-Language Models. (4%)Shihong Liu; Zhiqiu Lin; Samuel Yu; Ryan Lee; Tiffany Ling; Deepak Pathak; Deva Ramanan Unveiling Signle-Bit-Flip Attacks on DNN Executables. (1%)Yanzuo The Hong Kong University of Science and Technology Chen; Zhibo The Hong Kong University of Science and Technology Liu; Yuanyuan The Hong Kong University of Science and Technology Yuan; Sihang Huawei Technologies Hu; Tianxiang Huawei Technologies Li; Shuai The Hong Kong University of Science and Technology Wang 2023-09-11 Generalized Attacks on Face Verification Systems. (88%)Ehsan Nazari; Paula Branco; Guy-Vincent Jourdan Adversarial Attacks Assessment of Salient Object Detection via Symbolic Learning. (76%)Gustavo Olague; Roberto Pineda; Gerardo Ibarra-Vazquez; Matthieu Olague; Axel Martinez; Sambit Bakshi; Jonathan Vargas; Isnardo Reducindo Exploiting Machine Unlearning for Backdoor Attacks in Deep Learning System. (68%)Peixin Zhang; Jun Sun; Mingtian Tan; Xinyu Wang Privacy Side Channels in Machine Learning Systems. (10%)Edoardo Debenedetti; Giorgio Severi; Nicholas Carlini; Christopher A. Choquette-Choo; Matthew Jagielski; Milad Nasr; Eric Wallace; Florian Tramèr Divergences in Color Perception between Deep Neural Networks and Humans. (4%)Ethan O. Nadler; Elise Darragh-Ford; Bhargav Srinivasa Desikan; Christian Conaway; Mark Chu; Tasker Hull; Douglas Guilbeault Catch You Everything Everywhere: Guarding Textual Inversion via Concept Watermarking. (1%)Weitao Feng; Jiyan He; Jie Zhang; Tianwei Zhang; Wenbo Zhou; Weiming Zhang; Nenghai Yu Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs. (1%)Wenhua Cheng; Weiwei Zhang; Haihao Shen; Yiyang Cai; Xin He; Kaokao Lv 2023-09-10 Outlier Robust Adversarial Training. (98%)Shu Hu; Zhenhuan Yang; Xin Wang; Yiming Ying; Siwei Lyu DAD++: Improved Data-free Test Time Adversarial Defense. (98%)Gaurav Kumar Nayak; Inder Khatri; Shubham Randive; Ruchit Rawal; Anirban Chakraborty Machine Translation Models Stand Strong in the Face of Adversarial Attacks. (86%)Pavel Burnyshev; Elizaveta Kostenok; Alexey Zaytsev Secure Set-Based State Estimation for Linear Systems under Adversarial Attacks on Sensors. (3%)Muhammad Umar B. Niazi; Michelle S. Chong; Amr Alanwar; Karl H. Johansson 2023-09-09 Towards Robust Model Watermark via Reducing Parametric Vulnerability. (8%)Guanhao Gan; Yiming Li; Dongxian Wu; Shu-Tao Xia RecAD: Towards A Unified Library for Recommender Attack and Defense. (1%)Changsheng Wang; Jianbai Ye; Wenjie Wang; Chongming Gao; Fuli Feng; Xiangnan He 2023-09-08 Exploring Robust Features for Improving Adversarial Robustness. (99%)Hong Wang; Yuefan Deng; Shinjae Yoo; Yuewei Lin ARRTOC: Adversarially Robust Real-Time Optimization and Control. (2%)Akhil Ahmed; Rio-Chanona Ehecatl Antonio del; Mehmet Mercangoz Adversarial attacks on hybrid classical-quantum Deep Learning models for Histopathological Cancer Detection. (1%)Biswaraj Baral; Reek Majumdar; Bhavika Bhalgamiya; Taposh Dutta Roy Counterfactual Explanations via Locally-guided Sequential Algorithmic Recourse. (1%)Edward A. Small; Jeffrey N. Clark; Christopher J. McWilliams; Kacper Sokol; Jeffrey Chan; Flora D. Salim; Raul Santos-Rodriguez 2023-09-07 How adversarial attacks can disrupt seemingly stable accurate classifiers. (99%)Oliver J. Sutton; Qinghua Zhou; Ivan Y. Tyukin; Alexander N. Gorban; Alexander Bastounis; Desmond J. Higham Experimental Study of Adversarial Attacks on ML-based xApps in O-RAN. (99%)Naveen Naik Sapavath; Brian Kim; Kaushik Chowdhury; Vijay K Shah Adversarially Robust Deep Learning with Optimal-Transport-Regularized Divergences. (95%)Jeremiah Birrell; Mohammadreza Ebrahimi DiffDefense: Defending against Adversarial Attacks via Diffusion Models. (80%)Hondamunige Prasanna Silva; Lorenzo Seidenari; Bimbo Alberto Del One-to-Multiple Clean-Label Image Camouflage (OmClic) based Backdoor Attack on Deep Learning. (73%)Guohong Wang; Hua Ma; Yansong Gao; Alsharif Abuadbba; Zhi Zhang; Wei Kang; Said F. Al-Sarawib; Gongxuan Zhang; Derek Abbott Promoting Fairness in GNNs: A Characterization of Stability. (1%)Yaning Jia; Chunhui Zhang 2023-09-06 Certifying LLM Safety against Adversarial Prompting. (99%)Aounon Kumar; Chirag Agarwal; Suraj Srinivas; Aaron Jiaxun Li; Soheil Feizi; Himabindu Lakkaraju SWAP: Exploiting Second-Ranked Logits for Adversarial Attacks on Time Series. (84%)Chang George Dong; Liangwei Nathan Zheng; Weitong Chen; Wei Emma Zhang; Lin Yue Byzantine-Robust Federated Learning with Variance Reduction and Differential Privacy. (68%)Zikai Zhang; Rui Hu J-Guard: Journalism Guided Adversarially Robust Detection of AI-generated News. (38%)Tharindu Kumarage; Amrita Bhattacharjee; Djordje Padejski; Kristy Roschke; Dan Gillmor; Scott Ruston; Huan Liu; Joshua Garland MIRA: Cracking Black-box Watermarking on Deep Neural Networks via Model Inversion-based Removal Attacks. (22%)Yifan Lu; Wenxuan Li; Mi Zhang; Xudong Pan; Min Yang My Art My Choice: Adversarial Protection Against Unruly AI. (2%)Anthony Rhodes; Ram Bhagat; Umur Aybars Ciftci; Ilke Demir VeriDIP: Verifying Ownership of Deep Neural Networks through Privacy Leakage Fingerprints. (1%)Aoting Hu; Zhigang Lu; Renjie Xie; Minhui Xue A Theoretical Explanation of Activation Sparsity through Flat Minima and Adversarial Robustness. (1%)Ze Peng; Lei Qi; Yinghuan Shi; Yang Gao 2023-09-05 The Adversarial Implications of Variable-Time Inference. (99%)Dudi Biton; Aditi Misra; Efrat Levy; Jaidip Kotak; Ron Bitton; Roei Schuster; Nicolas Papernot; Yuval Elovici; Ben Nassi Adaptive Adversarial Training Does Not Increase Recourse Costs. (92%)Ian Hardy; Jayanth Yetukuri; Yang Liu Black-Box Attacks against Signed Graph Analysis via Balance Poisoning. (87%)Jialong Zhou; Yuni Lai; Jian Ren; Kai Zhou RobustEdge: Low Power Adversarial Detection for Cloud-Edge Systems. (83%)Abhishek Moitra; Abhiroop Bhattacharjee; Youngeun Kim; Priyadarshini Panda Building a Winning Team: Selecting Source Model Ensembles using a Submodular Transferability Estimation Approach. (4%)Vimal K B; Saketh Bachu; Tanmay Garg; Niveditha Lakshmi Narasimhan; Raghavan Konuru; Vineeth N Balasubramanian Robust Recommender System: A Survey and Future Directions. (2%)Kaike Zhang; Qi Cao; Fei Sun; Yunfan Wu; Shuchang Tao; Huawei Shen; Xueqi Cheng Dual Adversarial Alignment for Realistic Support-Query Shift Few-shot Learning. (1%)Siyang Jiang; Rui Fang; Hsi-Wen Chen; Wei Ding; Ming-Syan Chen 2023-09-04 Hindering Adversarial Attacks with Multiple Encrypted Patch Embeddings. (99%)AprilPyone MaungMaung; Isao Echizen; Hitoshi Kiya Improving Visual Quality and Transferability of Adversarial Attacks on Face Recognition Simultaneously with Adversarial Restoration. (99%)Fengfan Zhou; Hefei Ling; Yuxuan Shi; Jiazhong Chen; Ping Li Adv3D: Generating 3D Adversarial Examples in Driving Scenarios with NeRF. (99%)Leheng Li; Qing Lian; Ying-Cong Chen Toward Defensive Letter Design. (98%)Rentaro Kataoka; Akisato Kimura; Seiichi Uchida MathAttack: Attacking Large Language Models Towards Math Solving Ability. (97%)Zihao Zhou; Qiufeng Wang; Mingyu Jin; Jie Yao; Jianan Ye; Wei Liu; Wei Wang; Xiaowei Huang; Kaizhu Huang Efficient Defense Against Model Stealing Attacks on Convolutional Neural Networks. (93%)Kacem Khaled; Mouna Dhaouadi; Magalhães Felipe Gohring de; Gabriela Nicolescu Efficient Query-Based Attack against ML-Based Android Malware Detection under Zero Knowledge Setting. (92%)Ping He; Yifan Xia; Xuhong Zhang; Shouling Ji Safe and Robust Watermark Injection with a Single OoD Image. (8%)Shuyang Yu; Junyuan Hong; Haobo Zhang; Haotao Wang; Zhangyang Wang; Jiayu Zhou Dropout Attacks. (2%)Andrew Yuan; Alina Oprea; Cheng Tan Uncertainty in AI: Evaluating Deep Neural Networks on Out-of-Distribution Images. (2%)Jamiu Idowu; Ahmed Almasoud 2023-09-03 Robust and Efficient Interference Neural Networks for Defending Against Adversarial Attacks in ImageNet. (99%)Yunuo Xiong; Shujuan Liu; Hongwei Xiong Turn Fake into Real: Adversarial Head Turn Attacks Against Deepfake Detection. (98%)Weijie Wang; Zhengyu Zhao; Nicu Sebe; Bruno Lepri AdvMono3D: Advanced Monocular 3D Object Detection with Depth-Aware Robust Adversarial Training. (98%)Xingyuan Li; Jinyuan Liu; Long Ma; Xin Fan; Risheng Liu Robust Adversarial Defense by Tensor Factorization. (89%)Manish Bhattarai; Mehmet Cagri Kaymak; Ryan Barron; Ben Nebgen; Kim Rasmussen; Boian Alexandrov Dual Adversarial Resilience for Collaborating Robust Underwater Image Enhancement and Perception. (13%)Zengxi Zhang; Zhiying Jiang; Zeru Shi; Jinyuan Liu; Risheng Liu 2023-09-02 Towards Certified Probabilistic Robustness with High Accuracy. (98%)Ruihan Zhang; Peixin Zhang; Jun Sun Timbre-reserved Adversarial Attack in Speaker Identification. (98%)Qing Wang; Jixun Yao; Li Zhang; Pengcheng Guo; Lei Xie Regularly Truncated M-estimators for Learning with Noisy Labels. (1%)Xiaobo Xia; Pengqian Lu; Chen Gong; Bo Han; Jun Yu; Jun Yu; Tongliang Liu 2023-09-01 Baseline Defenses for Adversarial Attacks Against Aligned Language Models. (99%)Neel Jain; Avi Schwarzschild; Yuxin Wen; Gowthami Somepalli; John Kirchenbauer; Ping-yeh Chiang; Micah Goldblum; Aniruddha Saha; Jonas Geiping; Tom Goldstein Curating Naturally Adversarial Datasets for Trustworthy AI in Healthcare. (99%)Sydney Pugh; Ivan Ruchkin; Insup Lee; James Weimer Non-Asymptotic Bounds for Adversarial Excess Risk under Misspecified Models. (89%)Changyu Liu; Yuling Jiao; Junhui Wang; Jian Huang Why do universal adversarial attacks work on large language models?: Geometry might be the answer. (83%)Varshini Subhash; Anna Bialas; Weiwei Pan; Finale Doshi-Velez RenAIssance: A Survey into AI Text-to-Image Generation in the Era of Large Model. (1%)Fengxiang Bie; Yibo Yang; Zhongzhu Zhou; Adam Ghanem; Minjia Zhang; Zhewei Yao; Xiaoxia Wu; Connor Holmes; Pareesa Golnari; David A. Clifton; Yuxiong He; Dacheng Tao; Shuaiwen Leon Song Learned Visual Features to Textual Explanations. (1%)Saeid Asgari Taghanaki; Aliasghar Khani; Amir Khasahmadi; Aditya Sanghi; Karl D. D. Willis; Ali Mahdavi-Amiri 2023-08-31 Adversarial Finetuning with Latent Representation Constraint to Mitigate Accuracy-Robustness Tradeoff. (98%)Satoshi Suzuki; Shin'ya Yamaguchi; Shoichiro Takeda; Sekitoshi Kanai; Naoki Makishima; Atsushi Ando; Ryo Masumura Image Hijacking: Adversarial Images can Control Generative Models at Runtime. (98%)Luke Bailey; Euan Ong; Stuart Russell; Scott Emmons The Power of MEME: Adversarial Malware Creation with Model-Based Reinforcement Learning. (93%)Maria Rigaki; Sebastian Garcia Fault Injection and Safe-Error Attack for Extraction of Embedded Neural Network Models. (75%)Kevin Hector; Pierre-Alain Moellic; Mathieu Dumont; Jean-Max Dutertre Everyone Can Attack: Repurpose Lossy Compression as a Natural Backdoor Attack. (75%)Sze Jue Yang; Quang Nguyen; Chee Seng Chan; Khoa D. Doan FTA: Stealthy and Robust Backdoor Attack with Flexible Trigger on Federated Learning. (45%)Yanqi Qiao; Congwen Chen; Rui Wang; Kaitai Liang 2023-08-30 Explainable and Trustworthy Traffic Sign Detection for Safe Autonomous Driving: An Inductive Logic Programming Approach. (98%)Zahra University of Surrey Chaghazardi; Saber University of Surrey Fallah; Alireza University of Surrey Tamaddoni-Nezhad Robust Principles: Architectural Design Principles for Adversarially Robust CNNs. (11%)ShengYun Peng; Weilin Xu; Cory Cornelius; Matthew Hull; Kevin Li; Rahul Duggal; Mansi Phute; Jason Martin; Duen Horng Chau 2023-08-29 Adaptive Attack Detection in Text Classification: Leveraging Space Exploration Features for Text Sentiment Classification. (99%)Atefeh Mahdavi; Neda Keivandarian; Marco Carvalho Advancing Adversarial Robustness Through Adversarial Logit Update. (99%)Hao Xuan; Peican Zhu; Xingyu Li Imperceptible Adversarial Attack on Deep Neural Networks from Image Boundary. (99%)Fahad Alrasheedi; Xin Zhong A Classification-Guided Approach for Adversarial Attacks against Neural Machine Translation. (99%)Sahar Sadrizadeh; Ljiljana Dolamic; Pascal Frossard MDTD: A Multi Domain Trojan Detector for Deep Neural Networks. (97%)Arezoo Rajabi; Surudhi Asokraj; Fengqing Jiang; Luyao Niu; Bhaskar Ramasubramanian; Jim Ritcey; Radha Poovendran 3D Adversarial Augmentations for Robust Out-of-Domain Predictions. (87%)Alexander Lehner; Stefano Gasperini; Alvaro Marcos-Ramiro; Michael Schmidt; Nassir Navab; Benjamin Busam; Federico Tombari Everything Perturbed All at Once: Enabling Differentiable Graph Attacks. (84%)Haoran Liu; Bokun Wang; Jianling Wang; Xiangjue Dong; Tianbao Yang; James Caverlee Vulnerability of Machine Learning Approaches Applied in IoT-based Smart Grid: A Review. (75%)Zhenyong Zhang; Mengxiang Liu; Mingyang Sun; Ruilong Deng; Peng Cheng; Dusit Niyato; Mo-Yuen Chow; Jiming Chen Intriguing Properties of Diffusion Models: A Large-Scale Dataset for Evaluating Natural Attack Capability in Text-to-Image Generative Models. (67%)Takami Sato; Justin Yue; Nanze Chen; Ningfei Wang; Qi Alfred Chen Can We Rely on AI? (50%)Desmond J. Higham Uncertainty Aware Training to Improve Deep Learning Model Calibration for Classification of Cardiac MR Images. (1%)Tareen Dawood; Chen Chen; Baldeep S. Sidhua; Bram Ruijsink; Justin Goulda; Bradley Porter; Mark K. Elliott; Vishal Mehta; Christopher A. Rinaldi; Esther Puyol-Anton; Reza Razavi; Andrew P. King 2023-08-28 Adversarial Attacks on Foundational Vision Models. (80%)Nathan Inkawhich; Gwendolyn McDonald; Ryan Luley DiffSmooth: Certifiably Robust Learning via Diffusion Models and Local Smoothing. (45%)Jiawei Zhang; Zhongzhu Chen; Huan Zhang; Chaowei Xiao; Bo Li Identifying and Mitigating the Security Risks of Generative AI. (45%)Clark Barrett; Brad Boyd; Elie Burzstein; Nicholas Carlini; Brad Chen; Jihye Choi; Amrita Roy Chowdhury; Mihai Christodorescu; Anupam Datta; Soheil Feizi; Kathleen Fisher; Tatsunori Hashimoto; Dan Hendrycks; Somesh Jha; Daniel Kang; Florian Kerschbaum; Eric Mitchell; John Mitchell; Zulfikar Ramzan; Khawaja Shams; Dawn Song; Ankur Taly; Diyi Yang ReMAV: Reward Modeling of Autonomous Vehicles for Finding Likely Failure Events. (13%)Aizaz Sharif; Dusica Marijan Rep2wav: Noise Robust text-to-speech Using self-supervised representations. (1%)Qiushi Zhu; Yu Gu; Rilin Chen; Chao Weng; Yuchen Hu; Lirong Dai; Jie Zhang Are Existing Out-Of-Distribution Techniques Suitable for Network Intrusion Detection? (1%)Andrea Corsini; Shanchieh Jay Yang 2023-08-27 FaceChain: A Playground for Human-centric Artificial Intelligence Generated Content. (1%)Yang Liu; Cheng Yu; Lei Shang; Yongyi He; Ziheng Wu; Xingjun Wang; Chao Xu; Haoyu Xie; Weida Wang; Yuze Zhao; Lin Zhu; Chen Cheng; Weitao Chen; Yuan Yao; Wenmeng Zhou; Jiaqi Xu; Qiang Wang; Yingda Chen; Xuansong Xie; Baigui Sun Detecting Language Model Attacks with Perplexity. (1%)Gabriel Alon; Michael Kamfonas 2023-08-24 Exploring Transferability of Multimodal Adversarial Samples for Vision-Language Pre-training Models with Contrastive Learning. (99%)Youze Wang; Wenbo Hu; Yinpeng Dong; Richang Hong Don't Look into the Sun: Adversarial Solarization Attacks on Image Classifiers. (92%)Paul Gavrikov; Janis Keuper Evaluating the Vulnerabilities in ML systems in terms of adversarial attacks. (82%)John Harshith; Mantej Singh Gill; Madhan Jothimani Fast Adversarial Training with Smooth Convergence. (3%)Mengnan Zhao; Lihe Zhang; Yuqiu Kong; Baocai Yin WavMark: Watermarking for Audio Generation. (2%)Guangyu Chen; Yu Wu; Shujie Liu; Tao Liu; Xiaoyong Du; Furu Wei 2023-08-23 On-Manifold Projected Gradient Descent. (99%)Aaron Mahler; Tyrus Berry; Tom Stephens; Harbir Antil; Michael Merritt; Jeanie Schreiber; Ioannis Kevrekidis Sample Complexity of Robust Learning against Evasion Attacks. (98%)Pascale Gourdeau LCANets++: Robust Audio Classification using Multi-layer Neural Networks with Lateral Competition. (92%)Sayanton V. Dibbo; Juston S. Moore; Garrett T. Kenyon; Michael A. Teti BaDExpert: Extracting Backdoor Functionality for Accurate Backdoor Input Detection. (74%)Tinghao Xie; Xiangyu Qi; Ping He; Yiming Li; Jiachen T. Wang; Prateek Mittal RemovalNet: DNN Fingerprint Removal Attacks. (69%)Hongwei Yao; Zheng Li; Kunzhe Huang; Jian Lou; Zhan Qin; Kui Ren Graph Unlearning: A Review. (2%)Anwar Said; Tyler Derr; Mudassir Shabbir; Waseem Abbas; Xenofon Koutsoukos Ensembling Uncertainty Measures to Improve Safety of Black-Box Classifiers. (1%)Tommaso Zoppi; Andrea Ceccarelli; Andrea Bondavalli Aparecium: Revealing Secrets from Physical Photographs. (1%)Zhe Lei; Jie Zhang; Jingtao Li; Weiming Zhang; Nenghai Yu 2023-08-22 SEA: Shareable and Explainable Attribution for Query-based Black-box Attacks. (99%)Yue Gao; Ilia Shumailov; Kassem Fawaz Multi-Instance Adversarial Attack on GNN-Based Malicious Domain Detection. (99%)Mahmoud Nazzal; Issa Khalil; Abdallah Khreishah; NhatHai Phan; Yao Ma Does Physical Adversarial Example Really Matter to Autonomous Driving? Towards System-Level Effect of Adversarial Object Evasion Attack. (98%)Ningfei Wang; Yunpeng Luo; Takami Sato; Kaidi Xu; Qi Alfred Chen Protect Federated Learning Against Backdoor Attacks via Data-Free Trigger Generation. (86%)Yanxin Yang; Ming Hu; Yue Cao; Jun Xia; Yihao Huang; Yang Liu; Mingsong Chen Revisiting and Exploring Efficient Fast Adversarial Training via LAW: Lipschitz Regularization and Auto Weight Averaging. (76%)Xiaojun Jia; Yuefeng Chen; Xiaofeng Mao; Ranjie Duan; Jindong Gu; Rong Zhang; Hui Xue; Xiaochun Cao Designing an attack-defense game: how to increase robustness of financial transaction models via a competition. (75%)Alexey Zaytsev; Alex Natekin; Evgeni Vorsin; Valerii Smirnov; Oleg Sidorshin; Alexander Senin; Alexander Dudin; Dmitry Berestnev Adversarial Training Using Feedback Loops. (74%)Ali Haisam Muhammad Rafid; Adrian Sandu Adversarial Illusions in Multi-Modal Embeddings. (74%)Tingwei Zhang; Rishi Jha; Eugene Bagdasaryan; Vitaly Shmatikov LEAP: Efficient and Automated Test Method for NLP Software. (31%)Mingxuan Xiao; Yan Xiao; Hai Dong; Shunhui Ji; Pengcheng Zhang PatchBackdoor: Backdoor Attack against Deep Neural Networks without Model Modification. (16%)Yizhen Institute for AI Industry Research Yuan; Rui Shanghai Jiao Tong University, Shanghai, China Kong; Shenghao Wuhan University, Wuhan, China Xie; Yuanchun Institute for AI Industry Research Shanghai AI Laboratory, Shanghai, China Li; Yunxin Institute for AI Industry Research Shanghai AI Laboratory, Shanghai, China Liu 2023-08-21 Improving the Transferability of Adversarial Examples with Arbitrary Style Transfer. (99%)Zhijin Ge; Fanhua Shang; Hongying Liu; Yuanyuan Liu; Liang Wan; Wei Feng; Xiaosen Wang Spear and Shield: Adversarial Attacks and Defense Methods for Model-Based Link Prediction on Continuous-Time Dynamic Graphs. (99%)Dongjin Lee; Juho Lee; Kijung Shin Enhancing Adversarial Attacks: The Similar Target Method. (99%)Shuo Zhang; Ziruo Wang; Zikai Zhou; Huanran Chen Adversarial Attacks on Code Models with Discriminative Graph Patterns. (96%)Thanh-Dat Pick Nguyen; Yang Pick Zhou; Xuan Bach D. Pick Le; Pick Patanamon; Thongtanunam; David Lo Temporal-Distributed Backdoor Attack Against Video Based Action Recognition. (88%)Xi Li; Songhe Wang; Ruiquan Huang; Mahanth Gowda; George Kesidis Measuring the Effect of Causal Disentanglement on the Adversarial Robustness of Neural Network Models. (76%)Preben M. Ness; Dusica Marijan; Sunanda Bose Single-User Injection for Invisible Shilling Attack against Recommender Systems. (62%)Chengzhi Huang; Hui Li On the Adversarial Robustness of Multi-Modal Foundation Models. (4%)Christian Schlarmann; Matthias Hein Unlocking Accuracy and Fairness in Differentially Private Image Classification. (2%)Leonard Berrada; Soham De; Judy Hanwen Shen; Jamie Hayes; Robert Stanforth; David Stutz; Pushmeet Kohli; Samuel L. Smith; Borja Balle 2023-08-20 Boosting Adversarial Transferability by Block Shuffle and Rotation. (99%)Kunyu Wang; Xuanran He; Wenxuan Wang; Xiaosen Wang Improving Adversarial Robustness of Masked Autoencoders via Test-time Frequency-domain Prompting. (96%)Qidong Huang; Xiaoyi Dong; Dongdong Chen; Yinpeng Chen; Lu Yuan; Gang Hua; Weiming Zhang; Nenghai Yu HoSNN: Adversarially-Robust Homeostatic Spiking Neural Networks with Adaptive Firing Thresholds. (96%)Hejia Geng; Peng Li Hiding Backdoors within Event Sequence Data via Poisoning Attacks. (95%)Elizaveta Kovtun; Alina Ermilova; Dmitry Berestnev; Alexey Zaytsev Adversarial Collaborative Filtering for Free. (61%)Huiyuan Chen; Xiaoting Li; Vivian Lai; Chin-Chia Michael Yeh; Yujie Fan; Yan Zheng; Mahashweta Das; Hao Yang Efficient Joint Optimization of Layer-Adaptive Weight Pruning in Deep Neural Networks. (1%)Kaixin Xu; Zhe Wang; Xue Geng; Jie Lin; Min Wu; Xiaoli Li; Weisi Lin A Study on Robustness and Reliability of Large Language Model Code Generation. (1%)Li Zhong; Zilong Wang 2023-08-19 A Comparison of Adversarial Learning Techniques for Malware Detection. (99%)Pavla Louthánová; Matouš Kozák; Martin Jureček; Mark Stamp Robust Mixture-of-Expert Training for Convolutional Neural Networks. (83%)Yihua Zhang; Ruisi Cai; Tianlong Chen; Guanhua Zhang; Huan Zhang; Pin-Yu Chen; Shiyu Chang; Zhangyang Wang; Sijia Liu 2023-08-18 Black-box Adversarial Attacks against Dense Retrieval Models: A Multi-view Contrastive Learning Method. (99%)Yu-An Liu; Ruqing Zhang; Jiafeng Guo; Rijke Maarten de; Wei Chen; Yixing Fan; Xueqi Cheng Attacking logo-based phishing website detectors with adversarial perturbations. (99%)Jehyun Lee; Zhe Xin; Melanie Ng Pei See; Kanav Sabharwal; Giovanni Apruzzese; Dinil Mon Divakaran Compensating Removed Frequency Components: Thwarting Voice Spectrum Reduction Attacks. (92%)Shu Wang; Kun Sun; Qi Li DFB: A Data-Free, Low-Budget, and High-Efficacy Clean-Label Backdoor Attack. (54%)Binhao Ma; Jiahui Wang; Dejun Wang; Bo Meng Backdoor Mitigation by Correcting the Distribution of Neural Activations. (11%)Xi Li; Zhen Xiang; David J. Miller; George Kesidis On Gradient-like Explanation under a Black-box Setting: When Black-box Explanations Become as Good as White-box. (9%)Yi Cai; Gerhard Wunder Towards Attack-tolerant Federated Learning via Critical Parameter Analysis. (9%)Sungwon Han; Sungwon Park; Fangzhao Wu; Sundong Kim; Bin Zhu; Xing Xie; Meeyoung Cha Defending Label Inference Attacks in Split Learning under Regression Setting. (4%)Haoze Qiu; Fei Zheng; Chaochao Chen; Xiaolin Zheng An Image is Worth a Thousand Toxic Words: A Metamorphic Testing Framework for Content Moderation Software. (1%)Wenxuan Wang; Jingyuan Huang; Jen-tse Huang; Chang Chen; Jiazhen Gu; Pinjia He; Michael R. Lyu Proceedings of the 2nd International Workshop on Adaptive Cyber Defense. (1%)Marco Carvalho; Damian Marriott; Mark Bilinski; Ahmad Ridley 2023-08-17 AIR: Threats of Adversarial Attacks on Deep Learning-Based Information Recovery. (99%)Jinyin Chen; Jie Ge; Shilian Zheng; Linhui Ye; Haibin Zheng; Weiguo Shen; Keqiang Yue; Xiaoniu Yang Towards a Practical Defense against Adversarial Attacks on Deep Learning-based Malware Detectors via Randomized Smoothing. (99%)Daniel Gibert; Giulio Zizzo; Quan Le A White-Box False Positive Adversarial Attack Method on Contrastive Loss Based Offline Handwritten Signature Verification Models. (98%)Zhongliang Guo; Weiye Li; Yifei Qian; Ognjen Arandjelović; Lei Fang Causal Adversarial Perturbations for Individual Fairness and Robustness in Heterogeneous Data Spaces. (16%)Ahmad-Reza Ehyaei; Kiarash Mohammadi; Amir-Hossein Karimi; Samira Samadi; Golnoosh Farnadi That Doesn't Go There: Attacks on Shared State in Multi-User Augmented Reality Applications. (10%)Carter Slocum; Yicheng Zhang; Erfan Shayegani; Pedram Zaree; Nael Abu-Ghazaleh; Jiasi Chen Evaluating the Instruction-Following Robustness of Large Language Models to Prompt Injection. (10%)Zekun Li; Baolin Peng; Pengcheng He; Xifeng Yan General Lipschitz: Certified Robustness Against Resolvable Semantic Transformations via Transformation-Dependent Randomized Smoothing. (3%)Dmitrii Korzh; Mikhail Pautov; Olga Tsymboi; Ivan Oseledets 2023-08-16 Benchmarking Adversarial Robustness of Compressed Deep Learning Models. (81%)Brijesh Vora; Kartik Patwari; Syed Mahbub Hafiz; Zubair Shafiq; Chen-Nee Chuah Test-Time Poisoning Attacks Against Test-Time Adaptation Models. (73%)Tianshuo Cong; Xinlei He; Yun Shen; Yang Zhang Self-Deception: Reverse Penetrating the Semantic Firewall of Large Language Models. (67%)Zhenhua Wang; Wei Xie; Kai Chen; Baosheng Wang; Zhiwen Gui; Enze Wang Dynamic Neural Network is All You Need: Understanding the Robustness of Dynamic Mechanisms in Neural Networks. (61%)Mirazul Haque; Wei Yang Expressivity of Graph Neural Networks Through the Lens of Adversarial Robustness. (33%)Francesco Campi; Lukas Gosch; Tom Wollschläger; Yan Scholten; Stephan Günnemann 2023-08-15 SEDA: Self-Ensembling ViT with Defensive Distillation and Adversarial Training for robust Chest X-rays Classification. (99%)Raza Imam; Ibrahim Almakky; Salma Alrashdi; Baketah Alrashdi; Mohammad Yaqub Backpropagation Path Search On Adversarial Transferability. (99%)Zhuoer Xu; Zhangxuan Gu; Jianping Zhang; Shiwen Cui; Changhua Meng; Weiqiang Wang A Review of Adversarial Attacks in Computer Vision. (99%)Yutong Zhang; Yao Li; Yin Li; Zhichang Guo Robustness Over Time: Understanding Adversarial Examples' Effectiveness on Longitudinal Versions of Large Language Models. (95%)Yugeng Liu; Tianshuo Cong; Zhengyu Zhao; Michael Backes; Yun Shen; Yang Zhang Simple and Efficient Partial Graph Adversarial Attack: A New Perspective. (93%)Guanghui Zhu; Mengyu Chen; Chunfeng Yuan; Yihua Huang 2023-08-14 3DHacker: Spectrum-based Decision Boundary Generation for Hard-label 3D Point Cloud Attack. (99%)Yunbo Tao; Daizong Liu; Pan Zhou; Yulai Xie; Wei Du; Wei Hu White-Box Adversarial Attacks on Deep Learning-Based Radio Frequency Fingerprint Identification. (99%)Jie Ma; Junqing Zhang; Guanxiong Shen; Alan Marshall; Chip-Hong Chang AdvCLIP: Downstream-agnostic Adversarial Examples in Multimodal Contrastive Learning. (99%)Ziqi Zhou; Shengshan Hu; Minghui Li; Hangtao Zhang; Yechao Zhang; Hai Jin Enhancing the Antidote: Improved Pointwise Certifications against Poisoning Attacks. (68%)Shijie Liu; Andrew C. Cullen; Paul Montague; Sarah M. Erfani; Benjamin I. P. Rubinstein LLM Self Defense: By Self Examination, LLMs Know They Are Being Tricked. (54%)Alec Helbling; Mansi Phute; Matthew Hull; Duen Horng Chau DISBELIEVE: Distance Between Client Models is Very Essential for Effective Local Model Poisoning Attacks. (13%)Indu Joshi; Priyank Upadhya; Gaurav Kumar Nayak; Peter Schüffler; Nassir Navab ACTIVE: Towards Highly Transferable 3D Physical Camouflage for Universal and Robust Vehicle Evasion. (10%)Naufal Suryanto; Yongsu Kim; Harashta Tatimma Larasati; Hyoeun Kang; Thi-Thu-Huong Le; Yoonyoung Hong; Hunmin Yang; Se-Yoon Oh; Howon Kim SAM Meets Robotic Surgery: An Empirical Study on Generalization, Robustness and Adaptation. (1%)An Wang; Mobarakol Islam; Mengya Xu; Yang Zhang; Hongliang Ren 2023-08-13 SoK: Realistic Adversarial Attacks and Defenses for Intelligent Network Intrusion Detection. (99%)João Vitorino; Isabel Praça; Eva Maia Understanding the robustness difference between stochastic gradient descent and adaptive gradient methods. (45%)Avery Ma; Yangchen Pan; Amir-massoud Farahmand A Survey on Deep Neural Network Pruning-Taxonomy, Comparison, Analysis, and Recommendations. (1%)Hongrong Cheng; Miao Zhang; Javen Qinfeng Shi Robustified ANNs Reveal Wormholes Between Human Category Percepts. (1%)Guy Gaziv; Michael J. Lee; James J. DiCarlo Faithful to Whom? Questioning Interpretability Measures in NLP. (1%)Evan Crothers; Herna Viktor; Nathalie Japkowicz 2023-08-12 Not So Robust After All: Evaluating the Robustness of Deep Neural Networks to Unseen Adversarial Attacks. (99%)Roman Garaev; Bader Rasheed; Adil Khan One-bit Flip is All You Need: When Bit-flip Attack Meets Model Training. (13%)Jianshuo Dong; Han Qiu; Yiming Li; Tianwei Zhang; Yuanjie Li; Zeqi Lai; Chao Zhang; Shu-Tao Xia 2023-08-11 Enhancing Generalization of Universal Adversarial Perturbation through Gradient Aggregation. (98%)Xuannan Liu; Yaoyao Zhong; Yuhang Zhang; Lixiong Qin; Weihong Deng Physical Adversarial Attacks For Camera-based Smart Systems: Current Trends, Categorization, Applications, Research Challenges, and Future Outlook. (98%)Amira Guesmi; Muhammad Abdullah Hanif; Bassem Ouni; Muhammed Shafique Face Encryption via Frequency-Restricted Identity-Agnostic Attacks. (96%)Xin Dong; Rui Wang; Siyuan Liang; Aishan Liu; Lihua Jing White-box Membership Inference Attacks against Diffusion Models. (68%)Yan Pang; Tianhao Wang; Xuhui Kang; Mengdi Huai; Yang Zhang Test-Time Backdoor Defense via Detecting and Repairing. (10%)Jiyang Guan; Jian Liang; Ran He Continual Face Forgery Detection via Historical Distribution Preserving. (2%)Ke Sun; Shen Chen; Taiping Yao; Xiaoshuai Sun; Shouhong Ding; Rongrong Ji Fast and Accurate Transferability Measurement by Evaluating Intra-class Feature Variance. (1%)Huiwen Xu; U Kang 2023-08-10 Hard No-Box Adversarial Attack on Skeleton-Based Human Action Recognition with Skeleton-Motion-Informed Gradient. (99%)Zhengzhi Lu; He Wang; Ziyi Chang; Guoan Yang; Hubert P. H. Shum Symmetry Defense Against XGBoost Adversarial Perturbation Attacks. (96%)Blerta Lindqvist Complex Network Effects on the Robustness of Graph Convolutional Networks. (92%)Benjamin A. Miller; Kevin Chan; Tina Eliassi-Rad Critical Points ++: An Agile Point Cloud Importance Measure for Robust Classification, Adversarial Defense and Explainable AI. (80%)Meir Yossef Levi; Guy Gilboa State Machine Frameworks for Website Fingerprinting Defenses: Maybe Not. (61%)Ethan Witwer FLShield: A Validation Based Federated Learning Framework to Defend Against Poisoning Attacks. (45%)Ehsanul Kabir; Zeyu Song; Md Rafi Ur Rashid; Shagufta Mehnaz Comprehensive Analysis of Network Robustness Evaluation Based on Convolutional Neural Networks with Spatial Pyramid Pooling. (1%)Wenjun Jiang; Tianlong Fan; Changhao Li; Chuanfu Zhang; Tao Zhang; Zong-fu Luo 2023-08-09 Adv-Inpainting: Generating Natural and Transferable Adversarial Patch via Attention-guided Feature Fusion. (98%)Yanjie Li; Mingxing Duan; Bin Xiao Adversarial ModSecurity: Countering Adversarial SQL Injections with Robust Machine Learning. (93%)Biagio Montaruli; Luca Demetrio; Andrea Valenza; Battista Biggio; Luca Compagna; Davide Balzarotti; Davide Ariu; Luca Piras Adversarial Deep Reinforcement Learning for Cyber Security in Software Defined Networks. (81%)Luke Borchjes; Clement Nyirenda; Louise Leenen Data-Free Model Extraction Attacks in the Context of Object Detection. (41%)Harshit Shah; Aravindhan G; Pavan Kulkarni; Yuvaraj Govidarajulu; Manojkumar Parmar 2023-08-08 Pelta: Shielding Transformers to Mitigate Evasion Attacks in Federated Learning. (99%)Simon Queyrut; Yérom-David Bromberg; Valerio Schiavoni Federated Zeroth-Order Optimization using Trajectory-Informed Surrogate Gradients. (81%)Yao Shu; Xiaoqiang Lin; Zhongxiang Dai; Bryan Kian Hsiang Low The Model Inversion Eavesdropping Attack in Semantic Communication Systems. (67%)Yuhao Chen; Qianqian Yang; Zhiguo Shi; Jiming Chen Comprehensive Assessment of the Performance of Deep Learning Classifiers Reveals a Surprising Lack of Robustness. (64%)Michael W. Spratling XGBD: Explanation-Guided Graph Backdoor Detection. (54%)Zihan Guan; Mengnan Du; Ninghao Liu Improved Activation Clipping for Universal Backdoor Mitigation and Test-Time Detection. (50%)Hang Wang; Zhen Xiang; David J. Miller; George Kesidis Evil Operation: Breaking Speaker Recognition with PaddingBack. (31%)Zhe Ye; Diqun Yan; Li Dong; Kailai Shen Backdoor Federated Learning by Poisoning Backdoor-Critical Layers. (15%)Haomin Zhuang; Mingxian Yu; Hao Wang; Yang Hua; Jian Li; Xu Yuan 2023-08-07 Fixed Inter-Neuron Covariability Induces Adversarial Robustness. (98%)Muhammad Ahmed Shah; Bhiksha Raj Exploring the Physical World Adversarial Robustness of Vehicle Detection. (98%)Wei Jiang; Tianyuan Zhang; Shuangcheng Liu; Weiyu Ji; Zichao Zhang; Gang Xiao PAIF: Perception-Aware Infrared-Visible Image Fusion for Attack-Tolerant Semantic Segmentation. (86%)Zhu Liu; Jinyuan Liu; Benzhuang Zhang; Long Ma; Xin Fan; Risheng Liu A reading survey on adversarial machine learning: Adversarial attacks and their understanding. (81%)Shashank Kotyan A Four-Pronged Defense Against Byzantine Attacks in Federated Learning. (54%)Wei Wan; Shengshan Hu; Minghui Li; Jianrong Lu; Longling Zhang; Leo Yu Zhang; Hai Jin Improving Performance of Semi-Supervised Learning by Adversarial Attacks. (11%)Dongyoon Yang; Kunwoong Kim; Yongdai Kim Mondrian: Prompt Abstraction Attack Against Large Language Models for Cheaper API Pricing. (10%)Wai Man Si; Michael Backes; Yang Zhang 2023-08-06 SAAM: Stealthy Adversarial Attack on Monoculor Depth Estimation. (99%)Amira Guesmi; Muhammad Abdullah Hanif; Bassem Ouni; Muhammad Shafique CGBA: Curvature-aware Geometric Black-box Attack. (99%)Md Farhamdur Reza; Ali Rahmati; Tianfu Wu; Huaiyu Dai APBench: A Unified Benchmark for Availability Poisoning Attacks and Defenses. (98%)Tianrui Qin; Xitong Gao; Juanjuan Zhao; Kejiang Ye; Cheng-Zhong Xu Unsupervised Adversarial Detection without Extra Model: Training Loss Should Change. (82%)Chien Cheng Chyou; Hung-Ting Su; Winston H. Hsu Using Overlapping Methods to Counter Adversaries in Community Detection. (50%)Benjamin A. Miller; Kevin Chan; Tina Eliassi-Rad 2023-08-05 An Adaptive Model Ensemble Adversarial Attack for Boosting Adversarial Transferability. (99%)Bin Chen; Jia-Li Yin; Shukai Chen; Bo-Hao Chen; Ximeng Liu An AI-Enabled Framework to Defend Ingenious MDT-based Attacks on the Emerging Zero Touch Cellular Networks. (92%)Aneeqa Ijaz; Waseem Raza; Hasan Farooq; Marvin Manalastas; Ali Imran A Security and Usability Analysis of Local Attacks Against FIDO2. (1%)Tarun Kumar Yadav; Kent Seamons Approximating Positive Homogeneous Functions with Scale Invariant Neural Networks. (1%)Stefan Bamberger; Reinhard Heckel; Felix Krahmer 2023-08-04 Multi-attacks: Many images $+$ the same adversarial attack $\to$ many target labels. (99%)Stanislav Fort RobustMQ: Benchmarking Robustness of Quantized Models. (75%)Yisong Xiao; Aishan Liu; Tianyuan Zhang; Haotong Qin; Jinyang Guo; Xianglong Liu SureFED: Robust Federated Learning via Uncertainty-Aware Inward and Outward Inspection. (67%)Nasimeh Heydaribeni; Ruisi Zhang; Tara Javidi; Cristina Nita-Rotaru; Farinaz Koushanfar Vulnerabilities in AI Code Generators: Exploring Targeted Data Poisoning Attacks. (67%)Domenico Cotroneo; Cristina Improta; Pietro Liguori; Roberto Natella Universal Defensive Underpainting Patch: Making Your Text Invisible to Optical Character Recognition. (31%)JiaCheng Deng; Li Dong; Jiahao Chen; Diqun Yan; Rangding Wang; Dengpan Ye; Lingchen Zhao; Jinyu Tian BlindSage: Label Inference Attacks against Node-level Vertical Federated Graph Neural Networks. (9%)Marco Arazzi; Mauro Conti; Stefanos Koffas; Marina Krcek; Antonino Nocera; Stjepan Picek; Jing Xu 2023-08-03 Hard Adversarial Example Mining for Improving Robust Fairness. (99%)Chenhao Lin; Xiang Ji; Yulong Yang; Qian Li; Chao Shen; Run Wang; Liming Fang URET: Universal Robustness Evaluation Toolkit (for Evasion). (99%)Kevin Eykholt; Taesung Lee; Douglas Schales; Jiyong Jang; Ian Molloy; Masha Zorin AdvFAS: A robust face anti-spoofing framework against adversarial examples. (98%)Jiawei Chen; Xiao Yang; Heng Yin; Mingzhi Ma; Bihui Chen; Jianteng Peng; Yandong Guo; Zhaoxia Yin; Hang Su FROD: Robust Object Detection for Free. (67%)Muhammad; Awais; Weiming; Zhuang; Lingjuan; Lyu; Sung-Ho; Bae ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP. (33%)Lu Yan; Zhuo Zhang; Guanhong Tao; Kaiyuan Zhang; Xuan Chen; Guangyu Shen; Xiangyu Zhang From Prompt Injections to SQL Injection Attacks: How Protected is Your LLM-Integrated Web Application? (4%)Rodrigo Pedro; Daniel Castro; Paulo Carreira; Nuno Santos 2023-08-02 Inaudible Adversarial Perturbation: Manipulating the Recognition of User Speech in Real Time. (99%)Xinfeng Li; Chen Yan; Xuancun Lu; Zihan Zeng; Xiaoyu Ji; Wenyuan Xu Isolation and Induction: Training Robust Deep Neural Networks against Model Stealing Attacks. (98%)Jun Guo; Aishan Liu; Xingyu Zheng; Siyuan Liang; Yisong Xiao; Yichao Wu; Xianglong Liu Mercury: An Automated Remote Side-channel Attack to Nvidia Deep Learning Accelerator. (16%)Xiaobei Yan; Xiaoxuan Lou; Guowen Xu; Han Qiu; Shangwei Guo; Chip Hong Chang; Tianwei Zhang TEASMA: A Practical Approach for the Test Assessment of Deep Neural Networks using Mutation Analysis. (2%)Amin Abbasishahkoo; Mahboubeh Dadkhah; Lionel Briand; Dayi Lin LSF-IDM: Automotive Intrusion Detection Model with Lightweight Attribution and Semantic Fusion. (1%)Pengzhou Cheng; Lei Hua; Haobin Jiang; Mohammad Samie; Gongshen Liu 2023-08-01 Dynamic ensemble selection based on Deep Neural Network Uncertainty Estimation for Adversarial Robustness. (99%)Ruoxi Qin; Linyuan Wang; Xuehui Du; Xingyuan Chen; Bin Yan Improving Generalization of Adversarial Training via Robust Critical Fine-Tuning. (99%)Kaijie Zhu; Jindong Wang; Xixu Hu; Xing Xie; Ge Yang LimeAttack: Local Explainable Method for Textual Hard-Label Adversarial Attack. (99%)Hai Zhu; Zhaoqing Yang; Weiwei Shang; Yuren Wu Doubly Robust Instance-Reweighted Adversarial Training. (82%)Daouda Sow; Sen Lin; Zhangyang Wang; Yingbin Liang Training on Foveated Images Improves Robustness to Adversarial Attacks. (82%)Muhammad A. Shah; Bhiksha Raj Kidnapping Deep Learning-based Multirotors using Optimized Flying Adversarial Patches. (47%)Pia Hanfeld; Khaled Wahba; Marina M. -C. Höhne; Michael Bussmann; Wolfgang Hönig Robust Linear Regression: Phase-Transitions and Precise Tradeoffs for General Norms. (22%)Elvis Dohmatob; Meyer Scetbon Learning to Generate Training Datasets for Robust Semantic Segmentation. (9%)Marwane Hariat; Olivier Laurent; Rémi Kazmierczak; Shihao Zhang; Andrei Bursuc; Angela Yao; Gianni Franchi Zero-Shot Learning by Harnessing Adversarial Samples. (1%)Zhi Chen; Pengfei Zhang; Jingjing Li; Sen Wang; Zi Huang A Novel Cross-Perturbation for Single Domain Generalization. (1%)Dongjia Zhao; Lei Qi; Xiao Shi; Yinghuan Shi; Xin Geng 2023-07-31 A Novel Deep Learning based Model to Defend Network Intrusion Detection System against Adversarial Attacks. (99%)Khushnaseeb Roshan; Aasim Zafar; Shiekh Burhan Ul Haque Transferable Attack for Semantic Segmentation. (99%)Mengqi He; Jing Zhang; Zhaoyuan Yang; Mingyi He; Nick Barnes; Yuchao Dai Universal Adversarial Defense in Remote Sensing Based on Pre-trained Denoising Diffusion Models. (99%)Weikang Yu; Yonghao Xu; Pedram Ghamisi Defense of Adversarial Ranking Attack in Text Retrieval: Benchmark and Baseline via Detection. (97%)Xuanang Chen; Ben He; Le Sun; Yingfei Sun Text-CRS: A Generalized Certified Robustness Framework against Textual Adversarial Attacks. (86%)Xinyu Zhang; Hanbin Hong; Yuan Hong; Peng Huang; Binghui Wang; Zhongjie Ba; Kui Ren BAGM: A Backdoor Attack for Manipulating Text-to-Image Generative Models. (26%)Jordan Vice; Naveed Akhtar; Richard Hartley; Ajmal Mian Adversarially Robust Neural Legal Judgement Systems. (11%)Rohit Raj; V Susheela Devi Virtual Prompt Injection for Instruction-Tuned Large Language Models. (10%)Jun Yan; Vikas Yadav; Shiyang Li; Lichang Chen; Zheng Tang; Hai Wang; Vijay Srinivasan; Xiang Ren; Hongxia Jin Noisy Self-Training with Data Augmentations for Offensive and Hate Speech Detection Tasks. (1%)João A. Leite; Carolina Scarton; Diego F. Silva 2023-07-30 Theoretically Principled Trade-off for Stateful Defenses against Query-Based Black-Box Attacks. (99%)Ashish Hooda; Neal Mangaokar; Ryan Feng; Kassem Fawaz; Somesh Jha; Atul Prakash Benchmarking and Analyzing Robust Point Cloud Recognition: Bag of Tricks for Defending Adversarial Examples. (99%)Qiufan Ji; Lin Wang; Cong Shi; Shengshan Hu; Yingying Chen; Lichao Sun Probabilistically robust conformal prediction. (91%)Subhankar Ghosh; Yuanjie Shi; Taha Belkhouja; Yan Yan; Jana Doppa; Brian Jones On Updating Static Output Feedback Controllers Under State-Space Perturbation. (1%)MirSaleh Bahavarnia; Ahmad F. Taha 2023-07-29 You Can Backdoor Personalized Federated Learning. (92%)Tiandi Ye; Cen Chen; Yinggui Wang; Xiang Li; Ming Gao On Neural Network approximation of ideal adversarial attack and convergence of adversarial training. (92%)Rajdeep Haldar; Qifan Song Exposing Hidden Attackers in Industrial Control Systems using Micro-distortions. (41%)Suman Sourav; Binbin Chen 2023-07-28 Beating Backdoor Attack at Its Own Game. (97%)Min Liu; Alberto Sangiovanni-Vincentelli; Xiangyu Yue Adversarial training for tabular data with attack propagation. (67%)Tiago Leon Melo; João Bravo; Marco O. P. Sampaio; Paolo Romano; Hugo Ferreira; João Tiago Ascensão; Pedro Bizarro Improving Realistic Worst-Case Performance of NVCiM DNN Accelerators through Training with Right-Censored Gaussian Noise. (10%)Zheyu Yan; Yifan Qin; Wujie Wen; Xiaobo Sharon Hu; Yiyu Shi What can Discriminator do? Towards Box-free Ownership Verification of Generative Adversarial Network. (4%)Ziheng Huang; Boheng Li; Yan Cai; Run Wang; Shangwei Guo; Liming Fang; Jing Chen; Lina Wang 2023-07-27 R-LPIPS: An Adversarially Robust Perceptual Similarity Metric. (99%)Sara Ghazanfari; Siddharth Garg; Prashanth Krishnamurthy; Farshad Khorrami; Alexandre Araujo Universal and Transferable Adversarial Attacks on Aligned Language Models. (99%)Andy Zou; Zifan Wang; Nicholas Carlini; Milad Nasr; J. Zico Kolter; Matt Fredrikson When Measures are Unreliable: Imperceptible Adversarial Perturbations toward Top-$k$ Multi-Label Learning. (99%)Yuchen Sun; Qianqian Xu; Zitai Wang; Qingming Huang Backdoor Attacks for In-Context Learning with Language Models. (97%)Nikhil Kandpal; Matthew Jagielski; Florian Tramèr; Nicholas Carlini FLARE: Fingerprinting Deep Reinforcement Learning Agents using Universal Adversarial Masks. (93%)Buse G. A. Tekgul; N. Asokan Unified Adversarial Patch for Visible-Infrared Cross-modal Attacks in the Physical World. (92%)Xingxing Wei; Yao Huang; Yitong Sun; Jie Yu NSA: Naturalistic Support Artifact to Boost Network Confidence. (62%)Abhijith Sharma; Phil Munz; Apurva Narayan SEV-Step: A Single-Stepping Framework for AMD-SEV. (3%)Luca Wilke; Jan Wichelmann; Anja Rabich; Thomas Eisenbarth Decoding the Secrets of Machine Learning in Malware Classification: A Deep Dive into Datasets, Feature Extraction, and Model Performance. (1%)Savino Dambra; Yufei Han; Simone Aonzo; Platon Kotzias; Antonino Vitale; Juan Caballero; Davide Balzarotti; Leyla Bilge AC-Norm: Effective Tuning for Medical Image Analysis via Affine Collaborative Normalization. (1%)Chuyan Zhang; Yuncheng Yang; Hao Zheng; Yun Gu 2023-07-26 Enhanced Security against Adversarial Examples Using a Random Ensemble of Encrypted Vision Transformer Models. (99%)Ryota Iijima; Miki Tanaka; Sayaka Shiota; Hitoshi Kiya Set-level Guidance Attack: Boosting Adversarial Transferability of Vision-Language Pre-training Models. (99%)Dong Lu; Zhiqiang Wang; Teng Wang; Weili Guan; Hongchang Gao; Feng Zheng Defending Adversarial Patches via Joint Region Localizing and Inpainting. (99%)Junwen Chen; Xingxing Wei Lateral-Direction Localization Attack in High-Level Autonomous Driving: Domain-Specific Defense Opportunity via Lane Detection. (67%)Junjie Shen; Yunpeng Luo; Ziwen Wan; Qi Alfred Chen Plug and Pray: Exploiting off-the-shelf components of Multi-Modal Models. (33%)Erfan Shayegani; Yue Dong; Nael Abu-Ghazaleh Coupled-Space Attacks against Random-Walk-based Anomaly Detection. (11%)Yuni Lai; Marcin Waniek; Liying Li; Jingwen Wu; Yulin Zhu; Tomasz P. Michalak; Talal Rahwan; Kai Zhou FakeTracer: Proactively Defending Against Face-swap DeepFakes via Implanting Traces in Training. (5%)Pu Sun; Honggang Qi; Yuezun Li; Siwei Lyu Open Image Content Disarm And Reconstruction. (1%)Eli Belkind; Ran Dubin; Amit Dvir 2023-07-25 On the unreasonable vulnerability of transformers for image restoration -- and an easy fix. (99%)Shashank Agnihotri; Kanchana Vaishnavi Gandikota; Julia Grabinski; Paramanand Chandramouli; Margret Keuper Imperceptible Physical Attack against Face Recognition Systems via LED Illumination Modulation. (99%)Junbin Fang; Canjian Jiang; You Jiang; Puxi Lin; Zhaojie Chen; Yujing Sun; Siu-Ming Yiu; Zoe L. Jiang Foundational Models Defining a New Era in Vision: A Survey and Outlook. (10%)Muhammad Awais; Muzammal Naseer; Salman Khan; Rao Muhammad Anwer; Hisham Cholakkal; Mubarak Shah; Ming-Hsuan Yang; Fahad Shahbaz Khan Efficient Estimation of Average-Case Robustness for Multi-Class Classification. (10%)Tessa Han; Suraj Srinivas; Himabindu Lakkaraju 2023-07-24 Why Don't You Clean Your Glasses? Perception Attacks with Dynamic Optical Perturbations. (99%)Yi Han; Matthew Chan; Eric Wengrowski; Zhuohuan Li; Nils Ole Tippenhauer; Mani Srivastava; Saman Zonouz; Luis Garcia Lost In Translation: Generating Adversarial Examples Robust to Round-Trip Translation. (99%)Neel Bhandari; Pin-Yu Chen Data-free Black-box Attack based on Diffusion Model. (62%)Mingwen Shao; Lingzhuang Meng; Yuanjian Qiao; Lixu Zhang; Wangmeng Zuo Adaptive Certified Training: Towards Better Accuracy-Robustness Tradeoffs. (56%)Zhakshylyk Nurlanov; Frank R. Schmidt; Florian Bernard An Estimator for the Sensitivity to Perturbations of Deep Neural Networks. (31%)Naman Maheshwari; Nicholas Malaya; Scott Moe; Jaydeep P. Kulkarni; Sudhanva Gurumurthi Cyber Deception against Zero-day Attacks: A Game Theoretic Approach. (12%)Md Abu University of Texas at El Paso Sayed; Ahmed H. US Army Research Laboratory Anwar; Christopher University of Texas at El Paso Kiekintveld; Branislav Czech Technical University in Prague Bosansky; Charles US Army Research Laboratory Kamhoua Malware Resistant Data Protection in Hyper-connected Networks: A survey. (10%)Jannatul Ferdous; Rafiqul Islam; Maumita Bhattacharya; Md Zahidul Islam Investigating the Robustness of Sequential Recommender Systems Against Training Data Perturbations. (9%)Filippo Betello; Federico Siciliano; Pushkar Mishra; Fabrizio Silvestri Digital Twins for Moving Target Defense Validation in AC Microgrids. (1%)Suman Rath; Subham Sahoo; Shamik Sengupta Towards Bridging the FL Performance-Explainability Trade-Off: A Trustworthy 6G RAN Slicing Use-Case. (1%)Swastika Roy; Hatim Chergui; Christos Verikoukis Learning Provably Robust Estimators for Inverse Problems via Jittering. (1%)Anselm Krainovic; Mahdi Soltanolkotabi; Reinhard Heckel 2023-07-23 AdvDiff: Generating Unrestricted Adversarial Examples using Diffusion Models. (99%)Xuelong Dai; Kaisheng Liang; Bin Xiao Towards Generic and Controllable Attacks Against Object Detection. (99%)Guopeng Li; Yue Xu; Jian Ding; Gui-Song Xia Downstream-agnostic Adversarial Examples. (99%)Ziqi Zhou; Shengshan Hu; Ruizhi Zhao; Qian Wang; Leo Yu Zhang; Junhui Hou; Hai Jin Gradient-Based Word Substitution for Obstinate Adversarial Examples Generation in Language Models. (98%)Yimu Wang; Peng Shi; Hongyang Zhang A First Look at On-device Models in iOS Apps. (84%)Han Hu; Yujin Huang; Qiuyuan Chen; Terry Tue Zhuo; Chunyang Chen Robust Automatic Speech Recognition via WavAugment Guided Phoneme Adversarial Training. (83%)Gege Qi; Yuefeng Chen; Xiaofeng Mao; Xiaojun Jia; Ranjie Duan; Rong Zhang; Hui Xue Cross Contrastive Feature Perturbation for Domain Generalization. (1%)Chenming Li; Daoan Zhang; Wenjian Huang; Jianguo Zhang 2023-07-22 Backdoor Attacks against Voice Recognition Systems: A Survey. (13%)Baochen Yan; Jiahe Lan; Zheng Yan 2023-07-21 Unveiling Vulnerabilities in Interpretable Deep Learning Systems with Query-Efficient Black-box Attacks. (99%)Eldor Abdukhamidov; Mohammed Abuhamad; Simon S. Woo; Eric Chan-Tin; Tamer Abuhmed Fast Adaptive Test-Time Defense with Robust Features. (98%)Anurag Singh; Mahalakshmi Sabanayagam; Krikamol Muandet; Debarghya Ghoshdastidar FMT: Removing Backdoor Feature Maps via Feature Map Testing in Deep Neural Networks. (81%)Dong Huang; Qingwen Bu; Yahao Qing; Yichao Fu; Heming Cui Improving Viewpoint Robustness for Visual Recognition via Adversarial Training. (80%)Shouwei Ruan; Yinpeng Dong; Hang Su; Jianteng Peng; Ning Chen; Xingxing Wei OUTFOX: LLM-generated Essay Detection through In-context Learning with Adversarially Generated Examples. (62%)Ryuto Koike; Masahiro Kaneko; Naoaki Okazaki HybridAugment++: Unified Frequency Spectra Perturbations for Model Robustness. (26%)Mehmet Kerim Yucel; Ramazan Gokberk Cinbis; Pinar Duygulu Mitigating Communications Threats in Decentralized Federated Learning through Moving Target Defense. (1%)Enrique Tomás Martínez Beltrán; Pedro Miguel Sánchez Sánchez; Sergio López Bernal; Gérôme Bovet; Manuel Gil Pérez; Gregorio Martínez Pérez; Alberto Huertas Celdrán 2023-07-20 A LLM Assisted Exploitation of AI-Guardian. (98%)Nicholas Carlini Improving Transferability of Adversarial Examples via Bayesian Attacks. (98%)Qizhang Li; Yiwen Guo; Xiaochen Yang; Wangmeng Zuo; Hao Chen Adversarial attacks for mixtures of classifiers. (54%)Lucas Gnecco Heredia; Benjamin Negrevergne; Yann Chevaleyre PATROL: Privacy-Oriented Pruning for Collaborative Inference Against Model Inversion Attacks. (33%)Shiwei Ding; Lan Zhang; Miao Pan; Xiaoyong Yuan A Holistic Assessment of the Reliability of Machine Learning Systems. (4%)Anthony Corso; David Karamadian; Romeo Valentin; Mary Cooper; Mykel J. Kochenderfer Making Pre-trained Language Models both Task-solvers and Self-calibrators. (2%)Yangyi Chen; Xingyao Wang; Heng Ji Boundary State Generation for Testing and Improvement of Autonomous Driving Systems. (1%)Matteo Biagiola; Paolo Tonella A Survey of What to Share in Federated Learning: Perspectives on Model Utility, Privacy Leakage, and Communication Efficiency. (1%)Jiawei Shao; Zijian Li; Wenqiang Sun; Tailin Zhou; Yuchang Sun; Lumin Liu; Zehong Lin; Yuyi Mao; Jun Zhang 2023-07-19 Backdoor Attack against Object Detection with Clean Annotation. (93%)Yize Cheng; Wenbin Hu; Minhao Cheng Shared Adversarial Unlearning: Backdoor Mitigation by Unlearning Shared Adversarial Examples. (92%)Shaokui Wei; Mingda Zhang; Hongyuan Zha; Baoyuan Wu Rethinking Backdoor Attacks. (83%)Alaa Khaddaj; Guillaume Leclerc; Aleksandar Makelov; Kristian Georgiev; Hadi Salman; Andrew Ilyas; Aleksander Madry Towards Building More Robust Models with Frequency Bias. (81%)Qingwen Bu; Dong Huang; Heming Cui Reinforcing POD based model reduction techniques in reaction-diffusion complex networks using stochastic filtering and pattern recognition. (26%)Abhishek Ajayakumar; Soumyendu Raha 2023-07-18 CertPri: Certifiable Prioritization for Deep Neural Networks via Movement Cost in Feature Space. (67%)Haibin Zheng; Jinyin Chen; Haibo Jin FedDefender: Client-Side Attack-Tolerant Federated Learning. (50%)Sungwon Park; Sungwon Han; Fangzhao Wu; Sundong Kim; Bin Zhu; Xing Xie; Meeyoung Cha Can Neural Network Memorization Be Localized? (4%)Pratyush Maini; Michael C. Mozer; Hanie Sedghi; Zachary C. Lipton; J. Zico Kolter; Chiyuan Zhang 2023-07-17 Analyzing the Impact of Adversarial Examples on Explainable Machine Learning. (99%)Prathyusha Devabhakthini; Sasmita Parida; Raj Mani Shukla; Suvendu Chandan Nayak Adversarial Attacks on Traffic Sign Recognition: A Survey. (98%)Svetlana Pavlitska; Nico Lambing; J. Marius Zöllner Discretization-based ensemble model for robust learning in IoT. (87%)Anahita Namvar; Chandra Thapa; Salil S. Kanhere Unstoppable Attack: Label-Only Model Inversion via Conditional Diffusion Model. (83%)Rongke Liu; Dong Wang; Yizhi Ren; Zhen Wang; Kaitian Guo; Qianqian Qin; Xiaolei Liu Runtime Stealthy Perception Attacks against DNN-based Adaptive Cruise Control Systems. (22%)Xugui Zhou; Anqi Chen; Maxfield Kouzel; Haotian Ren; Morgan McCarty; Cristina Nita-Rotaru; Homa Alemzadeh On the Fly Neural Style Smoothing for Risk-Averse Domain Generalization. (2%)Akshay Mehra; Yunbei Zhang; Bhavya Kailkhura; Jihun Hamm A Machine Learning based Empirical Evaluation of Cyber Threat Actors High Level Attack Patterns over Low level Attack Patterns in Attributing Attacks. (1%)Umara Noor; Sawera Shahid; Rimsha Kanwal; Zahid Rashid 2023-07-16 Towards Viewpoint-Invariant Visual Recognition via Adversarial Training. (83%)Shouwei Ruan; Yinpeng Dong; Hang Su; Jianteng Peng; Ning Chen; Xingxing Wei Towards Stealthy Backdoor Attacks against Speech Recognition via Elements of Sound. (73%)Hanbo Cai; Pengcheng Zhang; Hai Dong; Yan Xiao; Stefanos Koffas; Yiming Li Diffusion to Confusion: Naturalistic Adversarial Patch Generation Based on Diffusion Model for Object Detector. (10%)Shuo-Yen Lin; Ernie Chu; Che-Hsien Lin; Jun-Cheng Chen; Jia-Ching Wang Lipschitz Continuous Algorithms for Covering Problems. (1%)Soh Kumabe; Yuichi Yoshida 2023-07-15 On the Robustness of Split Learning against Adversarial Attacks. (99%)Mingyuan Fan; Cen Chen; Chengyu Wang; Wenmeng Zhou; Jun Huang Why Does Little Robustness Help? Understanding and Improving Adversarial Transferability from Surrogate Training. (99%)Yechao Zhang; Shengshan Hu; Leo Yu Zhang; Junyu Shi; Minghui Li; Xiaogeng Liu; Wei Wan; Hai Jin Unified Adversarial Patch for Cross-modal Attacks in the Physical World. (92%)Xingxing Wei; Yao Huang; Yitong Sun; Jie Yu MasterKey: Automated Jailbreak Across Multiple Large Language Model Chatbots. (2%)Gelei Deng; Yi Liu; Yuekang Li; Kailong Wang; Ying Zhang; Zefeng Li; Haoyu Wang; Tianwei Zhang; Yang Liu 2023-07-14 Vulnerability-Aware Instance Reweighting For Adversarial Training. (99%)Olukorede Fakorede; Ashutosh Kumar Nirala; Modeste Atsague; Jin Tian Mitigating Adversarial Vulnerability through Causal Parameter Estimation by Adversarial Double Machine Learning. (99%)Byung-Kwan Lee; Junho Kim; Yong Man Ro On the Sensitivity of Deep Load Disaggregation to Adversarial Attacks. (99%)Hafsa Bousbiat; Yassine Himeur; Abbes Amira; Wathiq Mansoor RFLA: A Stealthy Reflected Light Adversarial Attack in the Physical World. (98%)Donghua Wang; Wen Yao; Tingsong Jiang; Chao Li; Xiaoqian Chen Frequency Domain Adversarial Training for Robust Volumetric Medical Segmentation. (98%)Asif Hanif; Muzammal Naseer; Salman Khan; Mubarak Shah; Fahad Shahbaz Khan Alleviating the Effect of Data Imbalance on Adversarial Training. (92%)Guanlin Li; Guowen Xu; Tianwei Zhang Structured Pruning of Neural Networks for Constraints Learning. (76%)Matteo Cacciola; Antonio Frangioni; Andrea Lodi Boosting Backdoor Attack with A Learnable Poisoning Sample Selection Strategy. (68%)Zihao Zhu; Mingda Zhang; Shaokui Wei; Li Shen; Yanbo Fan; Baoyuan Wu Erasing, Transforming, and Noising Defense Network for Occluded Person Re-Identification. (31%)Neng Dong; Liyan Zhang; Shuanglin Yan; Hao Tang; Jinhui Tang Omnipotent Adversarial Training in the Wild. (9%)Guanlin Li; Kangjie Chen; Yuan Xu; Han Qiu; Tianwei Zhang Certified Robustness for Large Language Models with Self-Denoising. (5%)Zhen Zhang; Guanhua Zhang; Bairu Hou; Wenqi Fan; Qing Li; Sijia Liu; Yang Zhang; Shiyu Chang 2023-07-13 Multi-objective Evolutionary Search of Variable-length Composite Semantic Perturbations. (99%)Jialiang Suna; Wen Yao; Tingsong Jianga; Xiaoqian Chena Introducing Foundation Models as Surrogate Models: Advancing Towards More Practical Adversarial Attacks. (99%)Jiaming Zhang; Jitao Sang; Qi Yi; Changsheng Xu Effective Prompt Extraction from Language Models. (4%)Yiming Zhang; Nicholas Carlini; Daphne Ippolito Layer-wise Linear Mode Connectivity. (1%)Linara Adilova; Maksym Andriushchenko; Michael Kamp; Asja Fischer; Martin Jaggi Defeating Proactive Jammers Using Deep Reinforcement Learning for Resource-Constrained IoT Networks. (1%)Abubakar Sani Ali; Shimaa Naser; Sami Muhaidat Towards Traitor Tracing in Black-and-White-Box DNN Watermarking with Tardos-based Codes. (1%)Elena Rodriguez-Lois; Fernando Perez-Gonzalez 2023-07-12 Single-Class Target-Specific Attack against Interpretable Deep Learning Systems. (99%)Eldor Abdukhamidov; Mohammed Abuhamad; George K. Thiruvathukal; Hyoungshick Kim; Tamer Abuhmed Microbial Genetic Algorithm-based Black-box Attack against Interpretable Deep Learning Systems. (99%)Eldor Abdukhamidov; Mohammed Abuhamad; Simon S. Woo; Eric Chan-Tin; Tamer Abuhmed Rational Neural Network Controllers. (2%)Matthew Newton; Antonis Papachristodoulou A Bayesian approach to quantifying uncertainties and improving generalizability in traffic prediction models. (1%)Agnimitra Sengupta; Sudeepta Mondal; Adway Das; S. Ilgin Guler Misclassification in Automated Content Analysis Causes Bias in Regression. Can We Fix It? Yes We Can! (1%)Nathan TeBlunthuis; Valerie Hase; Chung-Hong Chan 2023-07-11 ATWM: Defense against adversarial malware based on adversarial training. (99%)Kun Li; Fan Zhang; Wei Guo Membership Inference Attacks on DNNs using Adversarial Perturbations. (89%)Hassan Ali; Adnan Qayyum; Ala Al-Fuqaha; Junaid Qadir On the Vulnerability of DeepFake Detectors to Attacks Generated by Denoising Diffusion Models. (10%)Marija Ivanovska; Vitomir Štruc Random-Set Convolutional Neural Network (RS-CNN) for Epistemic Deep Learning. (4%)Shireen Kudukkil Manchingal; Muhammad Mubashar; Kaizheng Wang; Keivan Shariatmadar; Fabio Cuzzolin Differential Analysis of Triggers and Benign Features for Black-Box DNN Backdoor Detection. (2%)Hao Fu; Prashanth Krishnamurthy; Siddharth Garg; Farshad Khorrami Scale Alone Does not Improve Mechanistic Interpretability in Vision Models. (1%)Roland S. Zimmermann; Thomas Klein; Wieland Brendel Memorization Through the Lens of Curvature of Loss Function Around Samples. (1%)Isha Garg; Deepak Ravikumar; Kaushik Roy The Butterfly Effect in Artificial Intelligence Systems: Implications for AI Bias and Fairness. (1%)Emilio Ferrara 2023-07-10 Practical Trustworthiness Model for DNN in Dedicated 6G Application. (33%)Anouar Nechi; Ahmed Mahmoudi; Christoph Herold; Daniel Widmer; Thomas Kürner; Mladen Berekovic; Saleh Mulhem Distill-SODA: Distilling Self-Supervised Vision Transformer for Source-Free Open-Set Domain Adaptation in Computational Pathology. (1%)Guillaume Vray; Devavrat Tomar; Jean-Philippe Thiran; Behzad Bozorgtabar 2023-07-09 GNP Attack: Transferable Adversarial Examples via Gradient Norm Penalty. (98%)Tao Wu; Tie Luo; Donald C. Wunsch Enhancing Adversarial Robustness via Score-Based Optimization. (98%)Boya Zhang; Weijian Luo; Zhihua Zhang 2023-07-08 Adversarial Self-Attack Defense and Spatial-Temporal Relation Mining for Visible-Infrared Video Person Re-Identification. (99%)Huafeng Li; Le Xu; Yafei Zhang; Dapeng Tao; Zhengtao Yu Random Position Adversarial Patch for Vision Transformers. (83%)Mingzhen Shao Robust Ranking Explanations. (38%)Chao Chen; Chenghua Guo; Guixiang Ma; Ming Zeng; Xi Zhang; Sihong Xie 2023-07-07 A Theoretical Perspective on Subnetwork Contributions to Adversarial Robustness. (81%)Jovon Craig; Josh Andle; Theodore S. Nowak; Salimeh Yasaei Sekeh Fooling Contrastive Language-Image Pre-trained Models with CLIPMasterPrints. (68%)Matthias Freiberger; Peter Kun; Christian Igel; Anders Sundnes Løvlie; Sebastian Risi Scalable Membership Inference Attacks via Quantile Regression. (33%)Martin Bertran; Shuai Tang; Michael Kearns; Jamie Morgenstern; Aaron Roth; Zhiwei Steven Wu RADAR: Robust AI-Text Detection via Adversarial Learning. (5%)Xiaomeng Hu; Pin-Yu Chen; Tsung-Yi Ho Generation of Time-Varying Impedance Attacks Against Haptic Shared Control Steering Systems. (1%)Alireza Mohammadi; Hafiz Malik 2023-07-06 Sampling-based Fast Gradient Rescaling Method for Highly Transferable Adversarial Attacks. (99%)Xu Han; Anmin Liu; Chenxuan Yao; Yanbo Fan; Kun He NatLogAttack: A Framework for Attacking Natural Language Inference Models with Natural Logic. (92%)Zi'ou Zheng; Xiaodan Zhu Quantification of Uncertainty with Adversarial Models. (68%)Kajetan Schweighofer; Lukas Aichberger; Mykyta Ielanskyi; Günter Klambauer; Sepp Hochreiter A Vulnerability of Attribution Methods Using Pre-Softmax Scores. (41%)Miguel Lerma; Mirtha Lucas Probabilistic and Semantic Descriptions of Image Manifolds and Their Applications. (8%)Peter Tu; Zhaoyuan Yang; Richard Hartley; Zhiwei Xu; Jing Zhang; Yiwei Fu; Dylan Campbell; Jaskirat Singh; Tianyu Wang T-MARS: Improving Visual Representations by Circumventing Text Feature Learning. (1%)Pratyush Maini; Sachin Goyal; Zachary C. Lipton; J. Zico Kolter; Aditi Raghunathan 2023-07-05 Adversarial Attacks on Image Classification Models: FGSM and Patch Attacks and their Impact. (98%)Jaydip Sen; Subhasis Dasgupta DARE: Towards Robust Text Explanations in Biomedical and Healthcare Applications. (69%)Adam Ivankay; Mattia Rigotti; Pascal Frossard Detecting Images Generated by Deep Diffusion Models using their Local Intrinsic Dimensionality. (67%)Peter Lorenz; Ricard Durall; Janis Keuper GIT: Detecting Uncertainty, Out-Of-Distribution and Adversarial Samples using Gradients and Invariance Transformations. (62%)Julia Lust; Alexandru P. Condurache Securing Cloud FPGAs Against Power Side-Channel Attacks: A Case Study on Iterative AES. (5%)Nithyashankari Gummidipoondi JV Jayasankaran; Hao JV Guo; Satwik JV Patnaik; JV Jeyavijayan; Rajendran; Jiang Hu On the Adversarial Robustness of Generative Autoencoders in the Latent Space. (3%)Mingfei Lu; Badong Chen 2023-07-04 SCAT: Robust Self-supervised Contrastive Learning via Adversarial Training for Text Classification. (99%)Junjie Wu; Dit-Yan Yeung LEAT: Towards Robust Deepfake Disruption in Real-World Scenarios via Latent Ensemble Attack. (83%)Joonkyo Shim; Hyunsoo Yoon Interpretable Computer Vision Models through Adversarial Training: Unveiling the Robustness-Interpretability Connection. (68%)Delyan Boychev Overconfidence is a Dangerous Thing: Mitigating Membership Inference Attacks by Enforcing Less Confident Prediction. (45%)Zitao Chen; Karthik Pattabiraman Physically Realizable Natural-Looking Clothing Textures Evade Person Detectors via 3D Modeling. (26%)Zhanhao Hu; Wenda Chu; Xiaopei Zhu; Hui Zhang; Bo Zhang; Xiaolin Hu An Analysis of Untargeted Poisoning Attack and Defense Methods for Federated Online Learning to Rank Systems. (13%)Shuyi Wang; Guido Zuccon Machine Learning-Based Intrusion Detection: Feature Selection versus Feature Extraction. (1%)Vu-Duc Ngo; Tuan-Cuong Vuong; Luong Thien Van; Hung Tran Synthetic is all you need: removing the auxiliary data assumption for membership inference attacks against synthetic data. (1%)Florent Guépin; Matthieu Meeus; Ana-Maria Cretu; Montjoye Yves-Alexandre de 2023-07-03 Pareto-Secure Machine Learning (PSML): Fingerprinting and Securing Inference Serving Systems. (99%)Debopam Georgia Institute of Technology Sanyal; Jui-Tse Georgia Institute of Technology Hung; Manav Georgia Institute of Technology Agrawal; Prahlad Georgia Institute of Technology Jasti; Shahab University of California, Riverside Nikkhoo; Somesh University of Wisconsin-Madison Jha; Tianhao University of Virginia Wang; Sibin George Washington University Mohan; Alexey Georgia Institute of Technology Tumanov A Dual Stealthy Backdoor: From Both Spatial and Frequency Perspectives. (83%)Yudong Gao; Honglong Chen; Peng Sun; Junjian Li; Anqing Zhang; Zhibo Wang Analyzing the vulnerabilities in SplitFed Learning: Assessing the robustness against Data Poisoning Attacks. (62%)Aysha Thahsin Zahir Ismail; Raj Mani Shukla What Distributions are Robust to Indiscriminate Poisoning Attacks for Linear Learners? (62%)Fnu Suya; Xiao Zhang; Yuan Tian; David Evans Adversarial Learning in Real-World Fraud Detection: Challenges and Perspectives. (45%)Danele Lunghi; Alkis Simitsis; Olivier Caelen; Gianluca Bontempi Analysis of Task Transferability in Large Pre-trained Classifiers. (13%)Akshay Mehra; Yunbei Zhang; Jihun Hamm Enhancing the Robustness of QMIX against State-adversarial Attacks. (4%)Weiran Guo; Guanjun Liu; Ziyuan Zhou; Ling Wang; Jiacun Wang Towards Building Self-Aware Object Detectors via Reliable Uncertainty Quantification and Calibration. (1%)Kemal Oksuz; Tom Joy; Puneet K. Dokania 2023-07-02 Query-Efficient Decision-based Black-Box Patch Attack. (99%)Zhaoyu Chen; Bo Li; Shuang Wu; Shouhong Ding; Wenqiang Zhang Interpretability and Transparency-Driven Detection and Transformation of Textual Adversarial Examples (IT-DT). (99%)Bushra Sabir; M. Ali Babar; Sharif Abuadbba From ChatGPT to ThreatGPT: Impact of Generative AI in Cybersecurity and Privacy. (10%)Maanak Gupta; CharanKumar Akiri; Kshitiz Aryal; Eli Parker; Lopamudra Praharaj CLIMAX: An exploration of Classifier-Based Contrastive Explanations. (2%)Praharsh Nanavati; Ranjitha Prasad 2023-07-01 Common Knowledge Learning for Generating Transferable Adversarial Examples. (99%)Ruijie Yang; Yuanfang Guo; Junfu Wang; Jiantao Zhou; Yunhong Wang Adversarial Attacks and Defenses on 3D Point Cloud Classification: A Survey. (99%)Hanieh Naderi; Ivan V. Bajić Brightness-Restricted Adversarial Attack Patch. (75%)Mingzhen Shao Fedward: Flexible Federated Backdoor Defense Framework with Non-IID Data. (54%)Zekai Chen; Fuyi Wang; Zhiwei Zheng; Ximeng Liu; Yujie Lin Minimizing Energy Consumption of Deep Learning Models by Energy-Aware Training. (26%)Dario Lazzaro; Antonio Emanuele Cinà; Maura Pintor; Ambra Demontis; Battista Biggio; Fabio Roli; Marcello Pelillo SysNoise: Exploring and Benchmarking Training-Deployment System Inconsistency. (13%)Yan Wang; Yuhang Li; Ruihao Gong; Aishan Liu; Yanfei Wang; Jian Hu; Yongqiang Yao; Yunchen Zhang; Tianzi Xiao; Fengwei Yu; Xianglong Liu Gradients Look Alike: Sensitivity is Often Overestimated in DP-SGD. (10%)Anvith Thudi; Hengrui Jia; Casey Meehan; Ilia Shumailov; Nicolas Papernot CasTGAN: Cascaded Generative Adversarial Network for Realistic Tabular Data Synthesis. (5%)Abdallah Alshantti; Damiano Varagnolo; Adil Rasheed; Aria Rahmati; Frank Westad FedDefender: Backdoor Attack Defense in Federated Learning. (2%)Waris Virginia Tech Gill; Ali University of Minnesota Twin Cities Anwar; Muhammad Ali Virginia Tech Gulzar Hiding in Plain Sight: Differential Privacy Noise Exploitation for Evasion-resilient Localized Poisoning Attacks in Multiagent Reinforcement Learning. (1%)Md Tamjid Hossain; Hung La 2023-06-30 Defense against Adversarial Cloud Attack on Remote Sensing Salient Object Detection. (99%)Huiming Sun; Lan Fu; Jinlong Li; Qing Guo; Zibo Meng; Tianyun Zhang; Yuewei Lin; Hongkai Yu Efficient Backdoor Removal Through Natural Gradient Fine-tuning. (8%)Nazmul Karim; Abdullah Al Arafat; Umar Khalid; Zhishan Guo; Naznin Rahnavard Minimum-norm Sparse Perturbations for Opacity in Linear Systems. (1%)Varkey M John; Vaibhav Katewa 2023-06-29 Defending Black-box Classifiers by Bayesian Boundary Correction. (99%)He Wang; Yunfeng Diao Towards Optimal Randomized Strategies in Adversarial Example Game. (96%)Jiahao Xie; Chao Zhang; Weijie Liu; Wensong Bai; Hui Qian Neural Polarizer: A Lightweight and Effective Backdoor Defense via Purifying Poisoned Features. (13%)Mingli Zhu; Shaokui Wei; Hongyuan Zha; Baoyuan Wu NeuralFuse: Learning to Recover the Accuracy of Access-Limited Neural Network Inference in Low-Voltage Regimes. (1%)Hao-Lun Sun; Lei Hsiung; Nandhini Chandramoorthy; Pin-Yu Chen; Tsung-Yi Ho 2023-06-28 Mitigating the Accuracy-Robustness Trade-off via Multi-Teacher Adversarial Distillation. (99%)Shiji Zhao; Xizhe Wang; Xingxing Wei Boosting Adversarial Transferability with Learnable Patch-wise Masks. (99%)Xingxing Wei; Shiji Zhao Evaluating Similitude and Robustness of Deep Image Denoising Models via Adversarial Attack. (99%)Jie Ning; Yao Li; Zhichang Guo Group-based Robustness: A General Framework for Customized Robustness in the Real World. (98%)Weiran Lin; Keane Lucas; Neo Eyal; Lujo Bauer; Michael K. Reiter; Mahmood Sharif Distributional Modeling for Location-Aware Adversarial Patches. (98%)Xingxing Wei; Shouwei Ruan; Yinpeng Dong; Hang Su Enrollment-stage Backdoor Attacks on Speaker Recognition Systems via Adversarial Ultrasound. (98%)Xinfeng Li; Junning Ze; Chen Yan; Yushi Cheng; Xiaoyu Ji; Wenyuan Xu Does Saliency-Based Training bring Robustness for Deep Neural Networks in Image Classification? (93%)Ali Karkehabadi On Practical Aspects of Aggregation Defenses against Data Poisoning Attacks. (50%)Wenxiao Wang; Soheil Feizi On the Exploitability of Instruction Tuning. (13%)Manli Shu; Jiongxiao Wang; Chen Zhu; Jonas Geiping; Chaowei Xiao; Tom Goldstein 2023-06-27 Advancing Adversarial Training by Injecting Booster Signal. (98%)Hong Joo Lee; Youngjoon Yu; Yong Man Ro IMPOSITION: Implicit Backdoor Attack through Scenario Injection. (96%)Mozhgan Pourkeshavarz; Mohammad Sabokrou; Amir Rasouli Adversarial Training for Graph Neural Networks: Pitfalls, Solutions, and New Directions. (92%)Lukas Gosch; Simon Geisler; Daniel Sturm; Bertrand Charpentier; Daniel Zügner; Stephan Günnemann Robust Proxy: Improving Adversarial Robustness by Robust Proxy Learning. (89%)Hong Joo Lee; Yong Man Ro Your Attack Is Too DUMB: Formalizing Attacker Scenarios for Adversarial Transferability. (87%)Marco Alecci; Mauro Conti; Francesco Marchiori; Luca Martinelli; Luca Pajola [Re] Double Sampling Randomized Smoothing. (69%)Aryan Gupta; Sarthak Gupta; Abhay Kumar; Harsh Dugar Cooperation or Competition: Avoiding Player Domination for Multi-Target Robustness via Adaptive Budgets. (68%)Yimu Wang; Dinghuai Zhang; Yihan Wu; Heng Huang; Hongyang Zhang Catch Me If You Can: A New Low-Rate DDoS Attack Strategy Disguised by Feint. (26%)Tianyang Cai; Yuqi Li; Tao Jia; Leo Yu Zhang; Zheng Yang Shilling Black-box Review-based Recommender Systems through Fake Review Generation. (1%)Hung-Yun Chiang; Yi-Syuan Chen; Yun-Zhu Song; Hong-Han Shuai; Jason S. Chang 2023-06-26 On the Universal Adversarial Perturbations for Efficient Data-free Adversarial Detection. (99%)Songyang Gao; Shihan Dou; Qi Zhang; Xuanjing Huang; Jin Ma; Ying Shan Are aligned neural networks adversarially aligned? (99%)Nicholas Carlini; Milad Nasr; Christopher A. Choquette-Choo; Matthew Jagielski; Irena Gao; Anas Awadalla; Pang Wei Koh; Daphne Ippolito; Katherine Lee; Florian Tramer; Ludwig Schmidt The race to robustness: exploiting fragile models for urban camouflage and the imperative for machine learning security. (92%)Harriet Farlow; Matthew Garratt; Gavin Mount; Tim Lynar 3D-Aware Adversarial Makeup Generation for Facial Privacy Protection. (92%)Yueming Lyu; Yue Jiang; Ziwen He; Bo Peng; Yunfan Liu; Jing Dong Towards Sybil Resilience in Decentralized Learning. (80%)Thomas Werthenbach; Johan Pouwelse On the Resilience of Machine Learning-Based IDS for Automotive Networks. (78%)Ivo Zenden; Han Wang; Alfonso Iacovazzi; Arash Vahidi; Rolf Blom; Shahid Raza DSRM: Boost Textual Adversarial Training with Distribution Shift Risk Minimization. (75%)Songyang Gao; Shihan Dou; Yan Liu; Xiao Wang; Qi Zhang; Zhongyu Wei; Jin Ma; Ying Shan PWSHAP: A Path-Wise Explanation Model for Targeted Variables. (8%)Lucile Ter-Minassian; Oscar Clivio; Karla Diaz-Ordaz; Robin J. Evans; Chris Holmes 2023-06-25 A Spectral Perspective towards Understanding and Improving Adversarial Robustness. (99%)Binxiao Huang; Rui Lin; Chaofan Tao; Ngai Wong On Evaluating the Adversarial Robustness of Semantic Segmentation Models. (99%)Levente Halmosi; Mark Jelasity Robust Spatiotemporal Traffic Forecasting with Reinforced Dynamic Adversarial Training. (98%)Fan Liu; Weijia Zhang; Hao Liu Enhancing Adversarial Training via Reweighting Optimization Trajectory. (97%)Tianjin Huang; Shiwei Liu; Tianlong Chen; Meng Fang; Li Shen; Vlaod Menkovski; Lu Yin; Yulong Pei; Mykola Pechenizkiy RobuT: A Systematic Study of Table QA Robustness Against Human-Annotated Adversarial Perturbations. (87%)Yilun Zhao; Chen Zhao; Linyong Nan; Zhenting Qi; Wenlin Zhang; Xiangru Tang; Boyu Mi; Dragomir Radev Computational Asymmetries in Robust Classification. (80%)Samuele Marro; Michele Lombardi 2023-06-24 Boosting Model Inversion Attacks with Adversarial Examples. (98%)Shuai Zhou; Tianqing Zhu; Dayong Ye; Xin Yu; Wanlei Zhou Machine Learning needs Better Randomness Standards: Randomised Smoothing and PRNG-based attacks. (98%)Pranav Dahiya; Ilia Shumailov; Ross Anderson Similarity Preserving Adversarial Graph Contrastive Learning. (96%)Yeonjun In; Kanghoon Yoon; Chanyoung Park Weighted Automata Extraction and Explanation of Recurrent Neural Networks for Natural Language Tasks. (70%)Zeming Wei; Xiyue Zhang; Yihao Zhang; Meng Sun 2023-06-23 Creating Valid Adversarial Examples of Malware. (99%)Matouš Kozák; Martin Jureček; Mark Stamp; Troia Fabio Di Adversarial Robustness Certification for Bayesian Neural Networks. (92%)Matthew Wicker; Andrea Patane; Luca Laurenti; Marta Kwiatkowska A First Order Meta Stackelberg Method for Robust Federated Learning. (10%)Yunian Pan; Tao Li; Henger Li; Tianyi Xu; Zizhan Zheng; Quanyan Zhu 2023-06-22 Visual Adversarial Examples Jailbreak Large Language Models. (99%)Xiangyu Qi; Kaixuan Huang; Ashwinee Panda; Mengdi Wang; Prateek Mittal Towards quantum enhanced adversarial robustness in machine learning. (99%)Maxwell T. West; Shu-Lok Tsang; Jia S. Low; Charles D. Hill; Christopher Leckie; Lloyd C. L. Hollenberg; Sarah M. Erfani; Muhammad Usman Rethinking the Backward Propagation for Adversarial Transferability. (99%)Xiaosen Wang; Kangheng Tong; Kun He Evading Forensic Classifiers with Attribute-Conditioned Adversarial Faces. (96%)Fahad Shamshad; Koushik Srivatsan; Karthik Nandakumar Adversarial Resilience in Sequential Prediction via Abstention. (93%)Surbhi Goel; Steve Hanneke; Shay Moran; Abhishek Shetty Document Image Cleaning using Budget-Aware Black-Box Approximation. (92%)Ganesh Tata; Katyani Singh; Oeveren Eric Van; Nilanjan Ray Anticipatory Thinking Challenges in Open Worlds: Risk Management. (81%)Adam Amos-Binks; Dustin Dannenhauer; Leilani H. Gilpin Robust Semantic Segmentation: Strong Adversarial Attacks and Fast Training of Robust Models. (75%)Francesco Croce; Naman D Singh; Matthias Hein A First Order Meta Stackelberg Method for Robust Federated Learning (Technical Report). (33%)Henger Li; Tianyi Xu; Tao Li; Yunian Pan; Quanyan Zhu; Zizhan Zheng Impacts and Risk of Generative AI Technology on Cyber Defense. (4%)Subash Neupane; Ivan A. Fernandez; Sudip Mittal; Shahram Rahimi 2023-06-21 Adversarial Attacks Neutralization via Data Set Randomization. (99%)Mouna Rabhi; Pietro Roberto Di A Comprehensive Study on the Robustness of Image Classification and Object Detection in Remote Sensing: Surveying and Benchmarking. (92%)Shaohui Mei; Jiawei Lian; Xiaofei Wang; Yuru Su; Mingyang Ma; Lap-Pui Chau Sample Attackability in Natural Language Adversarial Attacks. (92%)Vyas Raina; Mark Gales Revisiting Image Classifier Training for Improved Certified Robust Defense against Adversarial Patches. (76%)Aniruddha Saha; Shuhua Yu; Arash Norouzzadeh; Wan-Yi Lin; Chaithanya Kumar Mummadi DP-BREM: Differentially-Private and Byzantine-Robust Federated Learning with Client Momentum. (47%)Xiaolan Gu; Ming Li; Li Xiong FFCV: Accelerating Training by Removing Data Bottlenecks. (3%)Guillaume Leclerc; Andrew Ilyas; Logan Engstrom; Sung Min Park; Hadi Salman; Aleksander Madry 2023-06-20 Reversible Adversarial Examples with Beam Search Attack and Grayscale Invariance. (99%)Haodong Zhang; Chi Man Pun; Xia Du Universal adversarial perturbations for multiple classification tasks with quantum classifiers. (99%)Yun-Zhong Qiu Physics-constrained Attack against Convolution-based Human Motion Prediction. (99%)Chengxu Duan; Zhicheng Zhang; Xiaoli Liu; Yonghao Dang; Jianqin Yin FDInet: Protecting against DNN Model Extraction via Feature Distortion Index. (50%)Hongwei Yao; Zheng Li; Haiqin Weng; Feng Xue; Kui Ren; Zhan Qin DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models. (33%)Boxin Wang; Weixin Chen; Hengzhi Pei; Chulin Xie; Mintong Kang; Chenhui Zhang; Chejian Xu; Zidi Xiong; Ritik Dutta; Rylan Schaeffer; Sang T. Truong; Simran Arora; Mantas Mazeika; Dan Hendrycks; Zinan Lin; Yu Cheng; Sanmi Koyejo; Dawn Song; Bo Li Towards a robust and reliable deep learning approach for detection of compact binary mergers in gravitational wave data. (3%)Shreejit Jadhav; Mihir Shrivastava; Sanjit Mitra Mitigating Speculation-based Attacks through Configurable Hardware/Software Co-design. (1%)Ali Hajiabadi; Archit Agarwal; Andreas Diavastos; Trevor E. Carlson LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical Imaging via Second-order Graph Matching. (1%)Duy M. H. Nguyen; Hoang Nguyen; Nghiem T. Diep; Tan N. Pham; Tri Cao; Binh T. Nguyen; Paul Swoboda; Nhat Ho; Shadi Albarqouni; Pengtao Xie; Daniel Sonntag; Mathias Niepert 2023-06-19 Comparative Evaluation of Recent Universal Adversarial Perturbations in Image Classification. (99%)Juanjuan Weng; Zhiming Luo; Dazhen Lin; Shaozi Li Adversarial Robustness of Prompt-based Few-Shot Learning for Natural Language Understanding. (75%)Venkata Prabhakara Sarath Nookala; Gaurav Verma; Subhabrata Mukherjee; Srijan Kumar Adversarial Training Should Be Cast as a Non-Zero-Sum Game. (73%)Alexander Robey; Fabian Latorre; George J. Pappas; Hamed Hassani; Volkan Cevher Eigenpatches -- Adversarial Patches from Principal Components. (38%)Jens Bayer; Stefan Becker; David Münch; Michael Arens Practical and General Backdoor Attacks against Vertical Federated Learning. (13%)Yuexin Xuan; Xiaojun Chen; Zhendong Zhao; Bisheng Tang; Ye Dong BNN-DP: Robustness Certification of Bayesian Neural Networks via Dynamic Programming. (5%)Steven Adams; Andrea Patane; Morteza Lahijanian; Luca Laurenti 2023-06-17 Edge Learning for 6G-enabled Internet of Things: A Comprehensive Survey of Vulnerabilities, Datasets, and Defenses. (98%)Mohamed Amine Ferrag; Othmane Friha; Burak Kantarci; Norbert Tihanyi; Lucas Cordeiro; Merouane Debbah; Djallel Hamouda; Muna Al-Hawawreh; Kim-Kwang Raymond Choo Understanding Certified Training with Interval Bound Propagation. (38%)Yuhao Mao; Mark Niklas Müller; Marc Fischer; Martin Vechev GlyphNet: Homoglyph domains dataset and detection using attention-based Convolutional Neural Networks. (9%)Akshat Gupta; Laxman Singh Tomar; Ridhima Garg Bkd-FedGNN: A Benchmark for Classification Backdoor Attacks on Federated Graph Neural Network. (1%)Fan Liu; Siqi Lai; Yansong Ning; Hao Liu 2023-06-16 Wasserstein distributional robustness of neural networks. (99%)Xingjian Bai; Guangyi He; Yifan Jiang; Jan Obloj Query-Free Evasion Attacks Against Machine Learning-Based Malware Detectors with Generative Adversarial Networks. (99%)Daniel Gibert; Jordi Planes; Quan Le; Giulio Zizzo You Don't Need Robust Machine Learning to Manage Adversarial Attack Risks. (98%)Edward Raff; Michel Benaroch; Andrew L. Farris Towards Better Certified Segmentation via Diffusion Models. (73%)Othmane Laousy; Alexandre Araujo; Guillaume Chassagnon; Marie-Pierre Revel; Siddharth Garg; Farshad Khorrami; Maria Vakalopoulou Adversarially robust clustering with optimality guarantees. (4%)Soham Jana; Kun Yang; Sanjeev Kulkarni CLIP2Protect: Protecting Facial Privacy using Text-Guided Makeup via Adversarial Latent Search. (1%)Fahad Shamshad; Muzammal Naseer; Karthik Nandakumar 2023-06-15 DIFFender: Diffusion-Based Adversarial Defense against Patch Attacks in the Physical World. (99%)Caixin Kang; Yinpeng Dong; Zhengyi Wang; Shouwei Ruan; Hang Su; Xingxing Wei OVLA: Neural Network Ownership Verification using Latent Watermarks. (64%)Feisi Fu; Wenchao Li Evaluating the Robustness of Text-to-image Diffusion Models against Real-world Attacks. (62%)Hongcheng Gao; Hao Zhang; Yinpeng Dong; Zhijie Deng On Strengthening and Defending Graph Reconstruction Attack with Markov Chain Approximation. (33%)Zhanke Zhou; Chenyu Zhou; Xuan Li; Jiangchao Yao; Quanming Yao; Bo Han Robustness Analysis on Foundational Segmentation Models. (9%)Madeline Chantry Schiappa; Sachidanand VS; Yunhao Ge; Ondrej Miksik; Yogesh S. Rawat; Vibhav Vineet Explore, Establish, Exploit: Red Teaming Language Models from Scratch. (1%)Stephen Casper; Jason Lin; Joe Kwon; Gatlen Culp; Dylan Hadfield-Menell Community Detection Attack against Collaborative Learning-based Recommender Systems. (1%)Yacine Belal; Sonia Ben Mokhtar; Mohamed Maouche; Anthony Simonet-Boulogne Concealing CAN Message Sequences to Prevent Schedule-based Bus-off Attacks. (1%)Sunandan Adhikary; Ipsita Koley; Arkaprava Sain; Soumyadeep das; Shuvam Saha; Soumyajit Dey 2023-06-14 Reliable Evaluation of Adversarial Transferability. (99%)Wenqian Yu; Jindong Gu; Zhijiang Li; Philip Torr A Relaxed Optimization Approach for Adversarial Attacks against Neural Machine Translation Models. (99%)Sahar Sadrizadeh; Clément Barbier; Ljiljana Dolamic; Pascal Frossard X-Detect: Explainable Adversarial Patch Detection for Object Detectors in Retail. (98%)Omer Hofman; Amit Giloni; Yarin Hayun; Ikuya Morikawa; Toshiya Shimizu; Yuval Elovici; Asaf Shabtai Augment then Smooth: Reconciling Differential Privacy with Certified Robustness. (98%)Jiapeng Wu; Atiyeh Ashari Ghomi; David Glukhov; Jesse C. Cresswell; Franziska Boenisch; Nicolas Papernot Efficient Backdoor Attacks for Deep Neural Networks in Real-world Scenarios. (83%)Hong Sun; Ziqiang Li; Pengfei Xia; Heng Li; Beihao Xia; Yi Wu; Bin Li A Unified Framework of Graph Information Bottleneck for Robustness and Membership Privacy. (75%)Enyan Dai; Limeng Cui; Zhengyang Wang; Xianfeng Tang; Yinghan Wang; Monica Cheng; Bing Yin; Suhang Wang On the Robustness of Latent Diffusion Models. (73%)Jianping Zhang; Zhuoer Xu; Shiwen Cui; Changhua Meng; Weibin Wu; Michael R. Lyu Improving Selective Visual Question Answering by Learning from Your Peers. (1%)Corentin Dancette; Spencer Whitehead; Rishabh Maheshwary; Ramakrishna Vedantam; Stefan Scherer; Xinlei Chen; Matthieu Cord; Marcus Rohrbach 2023-06-13 Theoretical Foundations of Adversarially Robust Learning. (99%)Omar Montasser Finite Gaussian Neurons: Defending against adversarial attacks by making neural networks say "I don't know". (99%)Felix Grezes I See Dead People: Gray-Box Adversarial Attack on Image-To-Text Models. (99%)Raz Lapid; Moshe Sipper Robustness of SAM: Segment Anything Under Corruptions and Beyond. (98%)Yu Qiao; Chaoning Zhang; Taegoo Kang; Donghun Kim; Chenshuang Zhang; Choong Seon Hong Area is all you need: repeatable elements make stronger adversarial attacks. (98%)Dillon Niederhut Malafide: a novel adversarial convolutive noise attack against deepfake and spoofing detection systems. (96%)Michele Panariello; Wanying Ge; Hemlata Tak; Massimiliano Todisco; Nicholas Evans Revisiting and Advancing Adversarial Training Through A Simple Baseline. (87%)Hong Liu Generative Watermarking Against Unauthorized Subject-Driven Image Synthesis. (78%)Yihan Ma; Zhengyu Zhao; Xinlei He; Zheng Li; Michael Backes; Yang Zhang Privacy Inference-Empowered Stealthy Backdoor Attack on Federated Learning under Non-IID Scenarios. (22%)Haochen Mei; Gaolei Li; Jun Wu; Longfei Zheng DHBE: Data-free Holistic Backdoor Erasing in Deep Neural Networks via Restricted Adversarial Distillation. (22%)Zhicong Yan; Shenghong Li; Ruijie Zhao; Yuan Tian; Yuanyuan Zhao Temporal Gradient Inversion Attacks with Robust Optimization. (8%)Bowen Li; Hanlin Gu; Ruoxin Chen; Jie Li; Chentao Wu; Na Ruan; Xueming Si; Lixin Fan Few-shot Multi-domain Knowledge Rearming for Context-aware Defence against Advanced Persistent Threats. (2%)Gaolei Li; Yuanyuan Zhao; Wenqi Wei; Yuchen Liu 2023-06-12 When Vision Fails: Text Attacks Against ViT and OCR. (99%)Nicholas Boucher; Jenny Blessing; Ilia Shumailov; Ross Anderson; Nicolas Papernot AROID: Improving Adversarial Robustness through Online Instance-wise Data Augmentation. (99%)Lin Li; Jianing Qiu; Michael Spratling How robust accuracy suffers from certified training with convex relaxations. (73%)Bartolomeis Piersilvio De; Jacob Clarysse; Amartya Sanyal; Fanny Yang Graph Agent Network: Empowering Nodes with Decentralized Communications Capabilities for Adversarial Resilience. (54%)Ao Liu; Wenshan Li; Tao Li; Beibei Li; Hanyuan Huang; Guangquan Xu; Pan Zhou Frequency-Based Vulnerability Analysis of Deep Learning Models against Image Corruptions. (13%)Harshitha Machiraju; Michael H. Herzog; Pascal Frossard On the Robustness of Removal-Based Feature Attributions. (11%)Chris Lin; Ian Covert; Su-In Lee VillanDiffusion: A Unified Backdoor Attack Framework for Diffusion Models. (1%)Sheng-Yen Chou; Pin-Yu Chen; Tsung-Yi Ho 2023-06-11 Securing Visually-Aware Recommender Systems: An Adversarial Image Reconstruction and Detection Framework. (99%)Minglei Yin; Bin Liu; Neil Zhenqiang Gong; Xin Li Neural Architecture Design and Robustness: A Dataset. (76%)Steffen Jung; Jovita Lukasik; Margret Keuper TrojLLM: A Black-box Trojan Prompt Attack on Large Language Models. (68%)Jiaqi Xue; Mengxin Zheng; Ting Hua; Yilin Shen; Yepeng Liu; Ladislau Boloni; Qian Lou 2023-06-10 Boosting Adversarial Robustness using Feature Level Stochastic Smoothing. (92%)Sravanti Addepalli; Samyak Jain; Gaurang Sriramanan; R. Venkatesh Babu NeRFool: Uncovering the Vulnerability of Generalizable Neural Radiance Fields against Adversarial Perturbations. (83%)Yonggan Fu; Ye Yuan; Souvik Kundu; Shang Wu; Shunyao Zhang; Yingyan Lin The Defense of Networked Targets in General Lotto games. (13%)Adel Aghajan; Keith Paarporn; Jason R. Marden 2023-06-09 Detecting Adversarial Directions in Deep Reinforcement Learning to Make Robust Decisions. (84%)Ezgi Korkmaz; Jonah Brown-Cohen When Authentication Is Not Enough: On the Security of Behavioral-Based Driver Authentication Systems. (70%)Emad Efatinasab; Francesco Marchiori; Denis Donadel; Alessandro Brighente; Mauro Conti Overcoming Adversarial Attacks for Human-in-the-Loop Applications. (45%)Ryan McCoppin; Marla Kennedy; Platon Lukyanenko; Sean Kennedy 2023-06-08 Adversarial Evasion Attacks Practicality in Networks: Testing the Impact of Dynamic Learning. (99%)Mohamed el Shehaby; Ashraf Matrawy Boosting Adversarial Transferability by Achieving Flat Local Maxima. (99%)Zhijin Ge; Hongying Liu; Xiaosen Wang; Fanhua Shang; Yuanyuan Liu COVER: A Heuristic Greedy Adversarial Attack on Prompt-based Learning in Language Models. (93%)Zihao Tan; Qingliang Chen; Wenbin Zhu; Yongjian Huang Generalizable Lightweight Proxy for Robust NAS against Diverse Perturbations. (83%)Hyeonjeong Ha; Minseon Kim; Sung Ju Hwang G$^2$uardFL: Safeguarding Federated Learning Against Backdoor Attacks through Attributed Client Graph Clustering. (62%)Hao Yu; Chuan Ma; Meng Liu; Xinwang Liu; Zhe Liu; Ming Ding A Melting Pot of Evolution and Learning. (41%)Moshe Sipper; Achiya Elyasaf; Tomer Halperin; Zvika Haramaty; Raz Lapid; Eyal Segal; Itai Tzruia; Snir Vitrack Tamam FedMLSecurity: A Benchmark for Attacks and Defenses in Federated Learning and LLMs. (13%)Shanshan Han; Baturalp Buyukates; Zijian Hu; Han Jin; Weizhao Jin; Lichao Sun; Xiaoyang Wang; Chulin Xie; Kai Zhang; Qifan Zhang; Yuhui Zhang; Chaoyang He; Salman Avestimehr PriSampler: Mitigating Property Inference of Diffusion Models. (12%)Hailong Hu; Jun Pang Investigating the Effect of Misalignment on Membership Privacy in the White-box Setting. (12%)Ana-Maria Cretu; Daniel Jones; Montjoye Yves-Alexandre de; Shruti Tople Robustness Testing for Multi-Agent Reinforcement Learning: State Perturbations on Critical Agents. (10%)Ziyuan Zhou; Guanjun Liu Conservative Prediction via Data-Driven Confidence Minimization. (8%)Caroline Choi; Fahim Tajwar; Yoonho Lee; Huaxiu Yao; Ananya Kumar; Chelsea Finn Robust Framework for Explanation Evaluation in Time Series Classification. (2%)Thu Trang Nguyen; Thach Le Nguyen; Georgiana Ifrim Enhancing Robustness of AI Offensive Code Generators via Data Augmentation. (2%)Cristina Improta; Pietro Liguori; Roberto Natella; Bojan Cukic; Domenico Cotroneo Open Set Relation Extraction via Unknown-Aware Training. (1%)Jun Zhao; Xin Zhao; Wenyu Zhan; Qi Zhang; Tao Gui; Zhongyu Wei; Yunwen Chen; Xiang Gao; Xuanjing Huang 2023-06-07 Extracting Cloud-based Model with Prior Knowledge. (99%)Shiqian Zhao; Kangjie Chen; Meng Hao; Jian Zhang; Guowen Xu; Hongwei Li; Tianwei Zhang Expanding Scope: Adapting English Adversarial Attacks to Chinese. (99%)Hanyu Liu; Chengyuan Cai; Yanjun Qi PromptAttack: Probing Dialogue State Trackers with Adversarial Prompts. (92%)Xiangjue Dong; Yun He; Ziwei Zhu; James Caverlee Optimal Transport Model Distributional Robustness. (83%)Van-Anh Nguyen; Trung Le; Anh Tuan Bui; Thanh-Toan Do; Dinh Phung PromptBench: Towards Evaluating the Robustness of Large Language Models on Adversarial Prompts. (76%)Kaijie Zhu; Jindong Wang; Jiaheng Zhou; Zichen Wang; Hao Chen; Yidong Wang; Linyi Yang; Wei Ye; Neil Zhenqiang Gong; Yue Zhang; Xing Xie A Linearly Convergent GAN Inversion-based Algorithm for Reverse Engineering of Deceptions. (45%)Darshan Thaker; Paris Giampouras; René Vidal Faithful Knowledge Distillation. (41%)Tom A. Lamb; Rudy Brunel; Krishnamurthy DJ Dvijotham; M. Pawan Kumar; Philip H. S. Torr; Francisco Eiras Divide and Repair: Using Options to Improve Performance of Imitation Learning Against Adversarial Demonstrations. (16%)Prithviraj Dasgupta Can current NLI systems handle German word order? Investigating language model performance on a new German challenge set of minimal pairs. (15%)Ines Reinig; Katja Markert Adversarial Sample Detection Through Neural Network Transport Dynamics. (10%)Skander Karkar; Patrick Gallinari; Alain Rakotomamonjy 2023-06-06 Revisiting the Trade-off between Accuracy and Robustness via Weight Distribution of Filters. (99%)Xingxing Wei; Shiji Zhao Avoid Adversarial Adaption in Federated Learning by Multi-Metric Investigations. (97%)Torsten University of Würzburg Krauß; Alexandra University of Würzburg Dmitrienko Transferable Adversarial Robustness for Categorical Data via Universal Robust Embeddings. (93%)Klim Kireev; Maksym Andriushchenko; Carmela Troncoso; Nicolas Flammarion Adversarial attacks and defenses in explainable artificial intelligence: A survey. (64%)Hubert Baniecki; Przemyslaw Biecek Exploring Model Dynamics for Accumulative Poisoning Discovery. (62%)Jianing Zhu; Xiawei Guo; Jiangchao Yao; Chao Du; Li He; Shuo Yuan; Tongliang Liu; Liang Wang; Bo Han Membership inference attack with relative decision boundary distance. (33%)JiaCheng Xu; ChengXiang Tan Performance-optimized deep neural networks are evolving into worse models of inferotemporal visual cortex. (8%)Drew Linsley; Ivan F. Rodriguez; Thomas Fel; Michael Arcaro; Saloni Sharma; Margaret Livingstone; Thomas Serre Adversarial Attacks and Defenses for Semantic Communication in Vehicular Metaverses. (1%)Jiawen Kang; Jiayi He; Hongyang Du; Zehui Xiong; Zhaohui Yang; Xumin Huang; Shengli Xie 2023-06-05 Adversarial alignment: Breaking the trade-off between the strength of an attack and its relevance to human perception. (99%)Drew Linsley; Pinyuan Feng; Thibaut Boissin; Alekh Karkada Ashok; Thomas Fel; Stephanie Olaiya; Thomas Serre Evading Black-box Classifiers Without Breaking Eggs. (99%)Edoardo Debenedetti; Nicholas Carlini; Florian Tramèr Evaluating robustness of support vector machines with the Lagrangian dual approach. (97%)Yuting Liu; Hong Gu; Pan Qin A Robust Likelihood Model for Novelty Detection. (93%)Ranya Almohsen; Shivang Patel; Donald A. Adjeroh; Gianfranco Doretto Adversarial Ink: Componentwise Backward Error Attacks on Deep Learning. (86%)Lucas Beerens; Desmond J. Higham Enhance Diffusion to Improve Robust Generalization. (76%)Jianhui Sun; Sanchit Sinha; Aidong Zhang KNOW How to Make Up Your Mind! Adversarially Detecting and Alleviating Inconsistencies in Natural Language Explanations. (68%)Myeongjun Jang; Bodhisattwa Prasad Majumder; Julian McAuley; Thomas Lukasiewicz; Oana-Maria Camburu Stable Diffusion is Unstable. (45%)Chengbin Du; Yanxi Li; Zhongwei Qiu; Chang Xu Neuron Activation Coverage: Rethinking Out-of-distribution Detection and Generalization. (1%)Yibing Liu; Chris Xing Tian; Haoliang Li; Lei Ma; Shiqi Wang Security Knowledge-Guided Fuzzing of Deep Learning Libraries. (1%)Nima Shiri Harzevili; Hung Viet Pham; Song Wang Input-gradient space particle inference for neural network ensembles. (1%)Trung Trinh; Markus Heinonen; Luigi Acerbi; Samuel Kaski 2023-06-04 Adversary for Social Good: Leveraging Adversarial Attacks to Protect Personal Attribute Privacy. (98%)Xiaoting Li; Lingwei Chen; Dinghao Wu Aerial Swarm Defense using Interception and Herding Strategies. (1%)Vishnu S. Chipade; Dimitra Panagou 2023-06-03 Towards Black-box Adversarial Example Detection: A Data Reconstruction-based Method. (99%)Yifei Gao; Zhiyu Lin; Yunfan Yang; Jitao Sang Learning to Defend by Attacking (and Vice-Versa): Transfer of Learning in Cybersecurity Games. (67%)Tyler Malloy; Cleotilde Gonzalez Can Directed Graph Neural Networks be Adversarially Robust? (56%)Zhichao Hou; Xitong Zhang; Wei Wang; Charu C. Aggarwal; Xiaorui Liu Flew Over Learning Trap: Learn Unlearnable Samples by Progressive Staged Training. (13%)Pucheng Dang; Xing Hu; Kaidi Xu; Jinhao Duan; Di Huang; Husheng Han; Rui Zhang; Zidong Du; Qi Guo; Yunji Chen Benchmarking Robustness of Adaptation Methods on Pre-trained Vision-Language Models. (1%)Shuo Chen; Jindong Gu; Zhen Han; Yunpu Ma; Philip Torr; Volker Tresp 2023-06-02 Towards Understanding Clean Generalization and Robust Overfitting in Adversarial Training. (99%)Binghui Li; Yuanzhi Li A Closer Look at the Adversarial Robustness of Deep Equilibrium Models. (92%)Zonghan Yang; Tianyu Pang; Yang Liu Adaptive Attractors: A Defense Strategy against ML Adversarial Collusion Attacks. (83%)Jiyi Zhang; Han Fang; Ee-Chien Chang Poisoning Network Flow Classifiers. (61%)Giorgio Severi; Simona Boboila; Alina Oprea; John Holodnak; Kendra Kratkiewicz; Jason Matterer Hyperparameter Learning under Data Poisoning: Analysis of the Influence of Regularization via Multiobjective Bilevel Optimization. (54%)Javier Carnerero-Cano; Luis Muñoz-González; Phillippa Spencer; Emil C. Lupu Robust low-rank training via approximate orthonormal constraints. (22%)Dayana Savostianova; Emanuele Zangrando; Gianluca Ceruti; Francesco Tudisco Supervised Adversarial Contrastive Learning for Emotion Recognition in Conversations. (13%)Dou Hu; Yinan Bao; Lingwei Wei; Wei Zhou; Songlin Hu Improving Adversarial Robustness of DEQs with Explicit Regulations Along the Neural Dynamics. (11%)Zonghan Yang; Peng Li; Tianyu Pang; Yang Liu Covert Communication Based on the Poisoning Attack in Federated Learning. (10%)Junchuan Liang; Rong Wang Invisible Image Watermarks Are Provably Removable Using Generative AI. (10%)Xuandong Zhao; Kexun Zhang; Zihao Su; Saastha Vasan; Ilya Grishchenko; Christopher Kruegel; Giovanni Vigna; Yu-Xiang Wang; Lei Li VoteTRANS: Detecting Adversarial Text without Training by Voting on Hard Labels of Transformations. (3%)Hoang-Quoc Nguyen-Son; Seira Hidano; Kazuhide Fukushima; Shinsaku Kiyomoto; Isao Echizen Unlearnable Examples for Diffusion Models: Protect Data from Unauthorized Exploitation. (2%)Zhengyue Zhao; Jinhao Duan; Xing Hu; Kaidi Xu; Chenan Wang; Rui Zhang; Zidong Du; Qi Guo; Yunji Chen MutateNN: Mutation Testing of Image Recognition Models Deployed on Hardware Accelerators. (1%)Nikolaos Louloudakis; Perry Gibson; José Cano; Ajitha Rajan Towards Robust GAN-generated Image Detection: a Multi-view Completion Representation. (1%)Chi Liu; Tianqing Zhu; Sheng Shen; Wanlei Zhou Improving the generalizability and robustness of large-scale traffic signal control. (1%)Tianyu Shi; Francois-Xavier Devailly; Denis Larocque; Laurent Charlin 2023-06-01 Adversarial Attack Based on Prediction-Correction. (99%)Chen Wan; Fangjun Huang Constructing Semantics-Aware Adversarial Examples with Probabilistic Perspective. (98%)Andi Zhang; Damon Wischik Reconstruction Distortion of Learned Image Compression with Imperceptible Perturbations. (96%)Yang Sui; Zhuohang Li; Ding Ding; Xiang Pan; Xiaozhong Xu; Shan Liu; Zhenzhong Chen Intriguing Properties of Text-guided Diffusion Models. (92%)Qihao Liu; Adam Kortylewski; Yutong Bai; Song Bai; Alan Yuille Robust Backdoor Attack with Visible, Semantic, Sample-Specific, and Compatible Triggers. (82%)Ruotong Wang; Hongrui Chen; Zihao Zhu; Li Liu; Yong Zhang; Yanbo Fan; Baoyuan Wu Improving the Robustness of Summarization Systems with Dual Augmentation. (76%)Xiuying Chen; Guodong Long; Chongyang Tao; Mingzhe Li; Xin Gao; Chengqi Zhang; Xiangliang Zhang Adversarial Robustness in Unsupervised Machine Learning: A Systematic Review. (38%)Mathias Lundteigen Mohus; Jinyue Li Does Black-box Attribute Inference Attacks on Graph Neural Networks Constitute Privacy Risk? (13%)Iyiola E. Olatunji; Anmar Hizber; Oliver Sihlovec; Megha Khosla CALICO: Self-Supervised Camera-LiDAR Contrastive Pre-training for BEV Perception. (13%)Jiachen Sun; Haizhong Zheng; Qingzhao Zhang; Atul Prakash; Z. Morley Mao; Chaowei Xiao ModelObfuscator: Obfuscating Model Information to Protect Deployed ML-based Systems. (4%)Mingyi Zhou; Xiang Gao; Jing Wu; John Grundy; Xiao Chen; Chunyang Chen; Li Li 2023-05-31 Exploring the Vulnerabilities of Machine Learning and Quantum Machine Learning to Adversarial Attacks using a Malware Dataset: A Comparative Analysis. (98%)Mst Shapna Akter; Hossain Shahriar; Iysa Iqbal; MD Hossain; M. A. Karim; Victor Clincy; Razvan Voicu Graph-based methods coupled with specific distributional distances for adversarial attack detection. (98%)Dwight Nwaigwe; Lucrezia Carboni; Martial Mermillod; Sophie Achard; Michel Dojat Adversarial-Aware Deep Learning System based on a Secondary Classical Machine Learning Verification Approach. (98%)Mohammed Alkhowaiter; Hisham Kholidy; Mnassar Alyami; Abdulmajeed Alghamdi; Cliff Zou Adversarial Clean Label Backdoor Attacks and Defenses on Text Classification Systems. (54%)Ashim Gupta; Amrith Krishna Deception by Omission: Using Adversarial Missingness to Poison Causal Structure Learning. (26%)Deniz Koyuncu; Alex Gittens; Bülent Yener; Moti Yung Red Teaming Language Model Detectors with Language Models. (15%)Zhouxing Shi; Yihan Wang; Fan Yin; Xiangning Chen; Kai-Wei Chang; Cho-Jui Hsieh Ambiguity in solving imaging inverse problems with deep learning based operators. (1%)Davide Evangelista; Elena Morotti; Elena Loli Piccolomini; James Nagy 2023-05-30 Pseudo-Siamese Network based Timbre-reserved Black-box Adversarial Attack in Speaker Identification. (99%)Qing Wang; Jixun Yao; Ziqian Wang; Pengcheng Guo; Lei Xie Breeding Machine Translations: Evolutionary approach to survive and thrive in the world of automated evaluation. (64%)Josef Jon; Ondřej Bojar Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness. (56%)Suraj Srinivas; Sebastian Bordt; Hima Lakkaraju Incremental Randomized Smoothing Certification. (33%)Shubham Ugare; Tarun Suresh; Debangshu Banerjee; Gagandeep Singh; Sasa Misailovic Defense Against Shortest Path Attacks. (16%)Benjamin A. Miller; Zohair Shafi; Wheeler Ruml; Yevgeniy Vorobeychik; Tina Eliassi-Rad; Scott Alfeld A Multilingual Evaluation of NER Robustness to Adversarial Inputs. (15%)Akshay Srinivasan; Sowmya Vajjala It begins with a boundary: A geometric view on probabilistically robust learning. (8%)Leon Bungert; Nicolás García Trillos; Matt Jacobs; Daniel McKenzie; Đorđe Nikolić; Qingsong Wang Adversarial Attacks on Online Learning to Rank with Stochastic Click Models. (2%)Zichen Wang; Rishab Balasubramanian; Hui Yuan; Chenyu Song; Mengdi Wang; Huazheng Wang Learning Perturbations to Explain Time Series Predictions. (1%)Joseph Enguehard 2023-05-29 From Adversarial Arms Race to Model-centric Evaluation: Motivating a Unified Automatic Robustness Evaluation Framework. (99%)Yangyi Chen; Hongcheng Gao; Ganqu Cui; Lifan Yuan; Dehan Kong; Hanlu Wu; Ning Shi; Bo Yuan; Longtao Huang; Hui Xue; Zhiyuan Liu; Maosong Sun; Heng Ji Fourier Analysis on Robustness of Graph Convolutional Neural Networks for Skeleton-based Action Recognition. (92%)Nariki Tanaka; Hiroshi Kera; Kazuhiko Kawamoto Exploiting Explainability to Design Adversarial Attacks and Evaluate Attack Resilience in Hate-Speech Detection Models. (92%)Pranath Reddy Kumbam; Sohaib Uddin Syed; Prashanth Thamminedi; Suhas Harish; Ian Perera; Bonnie J. Dorr UMD: Unsupervised Model Detection for X2X Backdoor Attacks. (81%)Zhen Xiang; Zidi Xiong; Bo Li Membership Inference Attacks against Language Models via Neighbourhood Comparison. (73%)Justus Mattern; Fatemehsadat Mireshghallah; Zhijing Jin; Bernhard Schölkopf; Mrinmaya Sachan; Taylor Berg-Kirkpatrick Trustworthy Sensor Fusion against Inaudible Command Attacks in Advanced Driver-Assistance System. (41%)Jiwei Guan; Lei Pan; Chen Wang; Shui Yu; Longxiang Gao; Xi Zheng Trainable and Explainable Simplicial Map Neural Networks. (41%)Eduardo Paluzo-Hidalgo; Miguel A. Gutiérrez-Naranjo; Rocio Gonzalez-Diaz Robust Lipschitz Bandits to Adversarial Corruptions. (11%)Yue Kang; Cho-Jui Hsieh; Thomas C. M. Lee Towards minimizing efforts for Morphing Attacks -- Deep embeddings for morphing pair selection and improved Morphing Attack Detection. (8%)Roman Kessler; Kiran Raja; Juan Tapia; Christoph Busch 2023-05-28 Amplification trojan network: Attack deep neural networks by amplifying their inherent weakness. (99%)Zhanhao Hu; Jun Zhu; Bo Zhang; Xiaolin Hu NaturalFinger: Generating Natural Fingerprint with Generative Adversarial Networks. (92%)Kang Yang; Kunhao Lai Backdoor Attacks Against Incremental Learners: An Empirical Evaluation Study. (41%)Yiqi Zhong; Xianming Liu; Deming Zhai; Junjun Jiang; Xiangyang Ji NOTABLE: Transferable Backdoor Attacks Against Prompt-based NLP Models. (38%)Kai Mei; Zheng Li; Zhenting Wang; Yang Zhang; Shiqing Ma Choose your Data Wisely: A Framework for Semantic Counterfactuals. (13%)Edmund Dervakos; Konstantinos Thomas; Giorgos Filandrianos; Giorgos Stamou BadLabel: A Robust Perspective on Evaluating and Enhancing Label-noise Learning. (5%)Jingfeng Zhang; Bo Song; Haohan Wang; Bo Han; Tongliang Liu; Lei Liu; Masashi Sugiyama Black-Box Anomaly Attribution. (1%)Tsuyoshi Idé; Naoki Abe 2023-05-27 Adversarial Attack On Yolov5 For Traffic And Road Sign Detection. (99%)Sanyam Jain Pre-trained transformer for adversarial purification. (99%)Kai Wu; Yujian Betterest Li; Xiaoyu Zhang; Handing Wang; Jing Liu Two Heads are Better than One: Towards Better Adversarial Robustness by Combining Transduction and Rejection. (98%)Nils Palumbo; Yang Guo; Xi Wu; Jiefeng Chen; Yingyu Liang; Somesh Jha Modeling Adversarial Attack on Pre-trained Language Models as Sequential Decision Making. (92%)Xuanjie Fang; Sijie Cheng; Yang Liu; Wei Wang On the Importance of Backbone to the Adversarial Robustness of Object Detectors. (83%)Xiao Li; Hang Chen; Xiaolin Hu No-Regret Online Reinforcement Learning with Adversarial Losses and Transitions. (2%)Tiancheng Jin; Junyan Liu; Chloé Rouyer; William Chang; Chen-Yu Wei; Haipeng Luo FoPro-KD: Fourier Prompted Effective Knowledge Distillation for Long-Tailed Medical Image Recognition. (1%)Marawan Elbatel; Robert Martí; Xiaomeng Li 2023-05-26 On Evaluating Adversarial Robustness of Large Vision-Language Models. (99%)Yunqing Zhao; Tianyu Pang; Chao Du; Xiao Yang; Chongxuan Li; Ngai-Man Cheung; Min Lin Leveraging characteristics of the output probability distribution for identifying adversarial audio examples. (98%)Matías P. Pizarro B.; Dorothea Kolossa; Asja Fischer Rethinking Adversarial Policies: A Generalized Attack Formulation and Provable Defense in Multi-Agent RL. (96%)Xiangyu Liu; Souradip Chakraborty; Yanchao Sun; Furong Huang A Tale of Two Approximations: Tightening Over-Approximation for DNN Robustness Verification via Under-Approximation. (45%)Zhiyi Xue; Si Liu; Zhaodi Zhang; Yiting Wu; Min Zhang Adversarial Attacks on Online Learning to Rank with Click Feedback. (38%)Jinhang Zuo; Zhiyao Zhang; Zhiyong Wang; Shuai Li; Mohammad Hajiesmaili; Adam Wierman DeepSeaNet: Improving Underwater Object Detection using EfficientDet. (2%)Sanyam Jain Trust-Aware Resilient Control and Coordination of Connected and Automated Vehicles. (1%)H M Sabbir Ahmad; Ehsan Sabouni; Wei Xiao; Christos G. Cassandras; Wenchao Li Efficient Detection of LLM-generated Texts with a Bayesian Surrogate Model. (1%)Zhijie Deng; Hongcheng Gao; Yibo Miao; Hao Zhang 2023-05-25 IDEA: Invariant Causal Defense for Graph Adversarial Robustness. (99%)Shuchang Tao; Qi Cao; Huawei Shen; Yunfan Wu; Bingbing Xu; Xueqi Cheng Don't Retrain, Just Rewrite: Countering Adversarial Perturbations by Rewriting Text. (98%)Ashim Gupta; Carter Wood Blum; Temma Choji; Yingjie Fei; Shalin Shah; Alakananda Vempala; Vivek Srikumar Diffusion-Based Adversarial Sample Generation for Improved Stealthiness and Controllability. (98%)Haotian Xue; Alexandre Araujo; Bin Hu; Yongxin Chen PEARL: Preprocessing Enhanced Adversarial Robust Learning of Image Deraining for Semantic Segmentation. (96%)Xianghao Jiao; Yaohua Liu; Jiaxin Gao; Xinyuan Chu; Risheng Liu; Xin Fan Adversarial Attacks on Leakage Detectors in Water Distribution Networks. (86%)Paul Stahlhofen; André Artelt; Luca Hermes; Barbara Hammer CARSO: Counter-Adversarial Recall of Synthetic Observations. (86%)Emanuele Ballarin; Alessio Ansuini; Luca Bortolussi On the Robustness of Segment Anything. (73%)Yihao Huang; Yue Cao; Tianlin Li; Felix Juefei-Xu; Di Lin; Ivor W. Tsang; Yang Liu; Qing Guo Detecting Adversarial Data by Probing Multiple Perturbations Using Expected Perturbation Score. (67%)Shuhai Zhang; Feng Liu; Jiahao Yang; Yifan Yang; Changsheng Li; Bo Han; Mingkui Tan Rethinking Diversity in Deep Neural Network Testing. (50%)Zi Wang; Jihye Choi; Ke Wang; Somesh Jha IMBERT: Making BERT Immune to Insertion-based Backdoor Attacks. (13%)Xuanli He; Jun Wang; Benjamin Rubinstein; Trevor Cohn Securing Deep Generative Models with Universal Adversarial Signature. (2%)Yu Zeng; Mo Zhou; Yuan Xue; Vishal M. Patel Concept-Centric Transformers: Enhancing Model Interpretability through Object-Centric Concept Learning within a Shared Global Workspace. (1%)Jinyung Hong; Keun Hee Park; Theodore P. Pavlic 2023-05-24 How do humans perceive adversarial text? A reality check on the validity and naturalness of word-based adversarial attacks. (99%)Salijona Dyrmishi; Salah Ghamizi; Maxime Cordy Robust Classification via a Single Diffusion Model. (99%)Huanran Chen; Yinpeng Dong; Zhengyi Wang; Xiao Yang; Chengqi Duan; Hang Su; Jun Zhu Introducing Competition to Boost the Transferability of Targeted Adversarial Examples through Clean Feature Mixup. (99%)Junyoung Byun; Myung-Joon Kwon; Seungju Cho; Yoonji Kim; Changick Kim Fantastic DNN Classifiers and How to Identify them without Data. (91%)Nathaniel Dean; Dilip Sarkar Adversarial Demonstration Attacks on Large Language Models. (88%)Jiongxiao Wang; Zichen Liu; Keun Hee Park; Muhao Chen; Chaowei Xiao Relating Implicit Bias and Adversarial Attacks through Intrinsic Dimension. (86%)Lorenzo Basile; Nikos Karantzas; Alberto D'Onofrio; Luca Bortolussi; Alex Rodriguez; Fabio Anselmi AdvFunMatch: When Consistent Teaching Meets Adversarial Robustness. (76%)Ziuhi Wu; Haichang Gao; Bingqian Zhou; Ping Wang Reconstructive Neuron Pruning for Backdoor Defense. (75%)Yige Li; Xixiang Lyu; Xingjun Ma; Nodens Koren; Lingjuan Lyu; Bo Li; Yu-Gang Jiang Another Dead End for Morphological Tags? Perturbed Inputs and Parsing. (74%)Alberto Muñoz-Ortiz; David Vilares Instructions as Backdoors: Backdoor Vulnerabilities of Instruction Tuning for Large Language Models. (50%)Jiashu Xu; Mingyu Derek Ma; Fei Wang; Chaowei Xiao; Muhao Chen From Shortcuts to Triggers: Backdoor Defense with Denoised PoE. (47%)Qin Liu; Fei Wang; Chaowei Xiao; Muhao Chen Clever Hans or Neural Theory of Mind? Stress Testing Social Reasoning in Large Language Models. (22%)Natalie Shapira; Mosh Levy; Seyed Hossein Alavi; Xuhui Zhou; Yejin Choi; Yoav Goldberg; Maarten Sap; Vered Shwartz Adversarial robustness of amortized Bayesian inference. (11%)Manuel Glöckler; Michael Deistler; Jakob H. Macke Sharpness-Aware Data Poisoning Attack. (10%)Pengfei He; Han Xu; Jie Ren; Yingqian Cui; Hui Liu; Charu C. Aggarwal; Jiliang Tang How to fix a broken confidence estimator: Evaluating post-hoc methods for selective classification with deep neural networks. (3%)Luís Felipe P. Cattelan; Danilo Silva M4: Multi-generator, Multi-domain, and Multi-lingual Black-Box Machine-Generated Text Detection. (1%)Yuxia Wang; Jonibek Mansurov; Petar Ivanov; Jinyan Su; Artem Shelmanov; Akim Tsvigun; Chenxi Whitehouse; Osama Mohammed Afzal; Tarek Mahmoud; Toru Sasaki; Thomas Arnold; Alham Fikri Aji; Nizar Habash; Iryna Gurevych; Preslav Nakov Ghostbuster: Detecting Text Ghostwritten by Large Language Models. (1%)Vivek Verma; Eve Fleisig; Nicholas Tomlin; Dan Klein 2023-05-23 The Best Defense is a Good Offense: Adversarial Augmentation against Adversarial Attacks. (99%)Iuri Frosio; Jan Kautz Enhancing Accuracy and Robustness through Adversarial Training in Class Incremental Continual Learning. (99%)Minchan Kwon; Kangil Kim QFA2SR: Query-Free Adversarial Transfer Attacks to Speaker Recognition Systems. (98%)Guangke Chen; Yedi Zhang; Zhe Zhao; Fu Song Expressive Losses for Verified Robustness via Convex Combinations. (95%)Palma Alessandro De; Rudy Bunel; Krishnamurthy Dvijotham; M. Pawan Kumar; Robert Stanforth; Alessio Lomuscio Impact of Light and Shadow on Robustness of Deep Neural Networks. (87%)Chengyin Hu; Weiwen Shi; Chao Li; Jialiang Sun; Donghua Wang; Junqi Wu; Guijian Tang A Causal View of Entity Bias in (Large) Language Models. (10%)Fei Wang; Wenjie Mo; Yiwei Wang; Wenxuan Zhou; Muhao Chen 2023-05-22 Latent Magic: An Investigation into Adversarial Examples Crafted in the Semantic Latent Space. (99%)BoYang Zheng Uncertainty-based Detection of Adversarial Attacks in Semantic Segmentation. (99%)Kira Maag; Asja Fischer FGAM:Fast Adversarial Malware Generation Method Based on Gradient Sign. (98%)Kun Li; Fan Zhang; Wei Guo Attribute-Guided Encryption with Facial Texture Masking. (98%)Chun Pong Lau; Jiang Liu; Rama Chellappa DiffProtect: Generate Adversarial Examples with Diffusion Models for Facial Privacy Protection. (98%)Jiang Liu; Chun Pong Lau; Rama Chellappa Byzantine Robust Cooperative Multi-Agent Reinforcement Learning as a Bayesian Game. (93%)Simin Li; Jun Guo; Jingqiao Xiu; Xini Yu; Jiakai Wang; Aishan Liu; Yaodong Yang; Xianglong Liu Towards Benchmarking and Assessing Visual Naturalness of Physical World Adversarial Attacks. (88%)Simin Li; Shuing Zhang; Gujun Chen; Dong Wang; Pu Feng; Jiakai Wang; Aishan Liu; Xin Yi; Xianglong Liu Flying Adversarial Patches: Manipulating the Behavior of Deep Learning-based Autonomous Multirotors. (54%)Pia Hanfeld; Marina M. -C. Höhne; Michael Bussmann; Wolfgang Hönig DeepBern-Nets: Taming the Complexity of Certifying Neural Networks using Bernstein Polynomial Activations and Precise Bound Propagation. (50%)Haitham Khedr; Yasser Shoukry The defender's perspective on automatic speaker verification: An overview. (22%)Haibin Wu; Jiawen Kang; Lingwei Meng; Helen Meng; Hung-yi Lee Model Stealing Attack against Multi-Exit Networks. (10%)Li Pan; Lv Peizhuo; Chen Kai; Cai Yuling; Xiang Fan; Zhang Shengzhi Adversarial Defenses via Vector Quantization. (8%)Zhiyi Dong; Yongyi Mao Adversarial Nibbler: A Data-Centric Challenge for Improving the Safety of Text-to-Image Models. (2%)Alicia Parrish; Hannah Rose Kirk; Jessica Quaye; Charvi Rastogi; Max Bartolo; Oana Inel; Juan Ciro; Rafael Mosquera; Addison Howard; Will Cukierski; D. Sculley; Vijay Janapa Reddi; Lora Aroyo Watermarking Classification Dataset for Copyright Protection. (1%)Yixin Liu; Hongsheng Hu; Xun Chen; Xuyun Zhang; Lichao Sun Improving Classifier Robustness through Active Generation of Pairwise Counterfactuals. (1%)Ananth Balashankar; Xuezhi Wang; Yao Qin; Ben Packer; Nithum Thain; Jilin Chen; Ed H. Chi; Alex Beutel Tied-Augment: Controlling Representation Similarity Improves Data Augmentation. (1%)Emirhan Kurtulus; Zichao Li; Yann Dauphin; Ekin Dogus Cubuk Adaptive Face Recognition Using Adversarial Information Network. (1%)Mei Wang; Weihong Deng 2023-05-21 Mist: Towards Improved Adversarial Examples for Diffusion Models. (99%)Chumeng Liang; Xiaoyu Wu Are Your Explanations Reliable? Investigating the Stability of LIME in Explaining Text Classifiers by Marrying XAI and Adversarial Attack. (81%)Christopher Burger; Lingwei Chen; Thai Le FAQ: Mitigating the Impact of Faults in the Weight Memory of DNN Accelerators through Fault-Aware Quantization. (1%)Muhammad Abdullah Hanif; Muhammad Shafique 2023-05-20 Dynamic Transformers Provide a False Sense of Efficiency. (92%)Yiming Chen; Simin Chen; Zexin Li; Wei Yang; Cong Liu; Robby T. Tan; Haizhou Li Annealing Self-Distillation Rectification Improves Adversarial Training. (76%)Yu-Yu Wu; Hung-Jui Wang; Shang-Tse Chen Stability, Generalization and Privacy: Precise Analysis for Random and NTK Features. (8%)Simone Bombari; Marco Mondelli 2023-05-19 Multi-Task Models Adversarial Attacks. (98%)Lijun Zhang; Xiao Liu; Kaleel Mahmood; Caiwen Ding; Hui Guan DAP: A Dynamic Adversarial Patch for Evading Person Detectors. (92%)Amira Guesmi; Ruitian Ding; Muhammad Abdullah Hanif; Ihsen Alouani; Muhammad Shafique Efficient ConvBN Blocks for Transfer Learning and Beyond. (67%)Kaichao You; Guo Qin; Anchang Bao; Meng Cao; Ping Huang; Jiulong Shan; Mingsheng Long Mitigating Backdoor Poisoning Attacks through the Lens of Spurious Correlation. (8%)Xuanli He; Qiongkai Xu; Jun Wang; Benjamin Rubinstein; Trevor Cohn Long-tailed Visual Recognition via Gaussian Clouded Logit Adjustment. (5%)Mengke Li; Yiu-ming Cheung; Yang Lu SneakyPrompt: Evaluating Robustness of Text-to-image Generative Models' Safety Filters. (4%)Yuchen Yang; Bo Hui; Haolin Yuan; Neil Gong; Yinzhi Cao Latent Imitator: Generating Natural Individual Discriminatory Instances for Black-Box Fairness Testing. (2%)Yisong Xiao; Aishan Liu; Tianlin Li; Xianglong Liu Controlling the Extraction of Memorized Data from Large Language Models via Prompt-Tuning. (1%)Mustafa Safa Ozdayi; Charith Peris; Jack FitzGerald; Christophe Dupuy; Jimit Majmudar; Haidar Khan; Rahil Parikh; Rahul Gupta 2023-05-18 Deep PackGen: A Deep Reinforcement Learning Framework for Adversarial Network Packet Generation. (99%)Soumyadeep Hore; Jalal Ghadermazi; Diwas Paudel; Ankit Shah; Tapas K. Das; Nathaniel D. Bastian Adversarial Amendment is the Only Force Capable of Transforming an Enemy into a Friend. (99%)Chong Yu; Tao Chen; Zhongxue Gan Architecture-agnostic Iterative Black-box Certified Defense against Adversarial Patches. (99%)Di Yang; Yihao Huang; Qing Guo; Felix Juefei-Xu; Ming Hu; Yang Liu; Geguang Pu Towards an Accurate and Secure Detector against Adversarial Perturbations. (99%)Chao Wang; Shuren Qi; Zhiqiu Huang; Yushu Zhang; Xiaochun Cao Quantifying the robustness of deep multispectral segmentation models against natural perturbations and data poisoning. (99%)Elise Bishoff; Charles Godfrey; Myles McKay; Eleanor Byler How Deep Learning Sees the World: A Survey on Adversarial Attacks & Defenses. (98%)Joana C. Costa; Tiago Roxo; Hugo Proença; Pedro R. M. Inácio RobustFair: Adversarial Evaluation through Fairness Confusion Directed Gradient Search. (93%)Xuran Li; Peng Wu; Kaixiang Dong; Zhen Zhang Attacks on Online Learners: a Teacher-Student Analysis. (54%)Riccardo Giuseppe Margiotta; Sebastian Goldt; Guido Sanguinetti Explaining V1 Properties with a Biologically Constrained Deep Learning Architecture. (47%)Galen Pogoncheff; Jacob Granley; Michael Beyeler Zero-Day Backdoor Attack against Text-to-Image Diffusion Models via Personalization. (2%)Yihao Huang; Qing Guo; Felix Juefei-Xu Large Language Models can be Guided to Evade AI-Generated Text Detection. (1%)Ning Lu; Shengcai Liu; Rui He; Ke Tang Re-thinking Data Availablity Attacks Against Deep Neural Networks. (1%)Bin Fang; Bo Li; Shuang Wu; Ran Yi; Shouhong Ding; Lizhuang Ma TrustSER: On the Trustworthiness of Fine-tuning Pre-trained Speech Embeddings For Speech Emotion Recognition. (1%)Tiantian Feng; Rajat Hebbar; Shrikanth Narayanan 2023-05-17 Content-based Unrestricted Adversarial Attack. (99%)Zhaoyu Chen; Bo Li; Shuang Wu; Kaixun Jiang; Shouhong Ding; Wenqiang Zhang Raising the Bar for Certified Adversarial Robustness with Diffusion Models. (95%)Thomas Altstidl; David Dobre; Björn Eskofier; Gauthier Gidel; Leo Schwinn The Adversarial Consistency of Surrogate Risks for Binary Classification. (10%)Natalie Frank; Jonathan Niles-Weed Variational Classification. (1%)Shehzaad Dhuliawala; Mrinmaya Sachan; Carl Allen Compress, Then Prompt: Improving Accuracy-Efficiency Trade-off of LLM Inference with Transferable Prompt. (1%)Zhaozhuo Xu; Zirui Liu; Beidi Chen; Yuxin Tang; Jue Wang; Kaixiong Zhou; Xia Hu; Anshumali Shrivastava PaLM 2 Technical Report. (1%)Rohan Anil; Andrew M. Dai; Orhan Firat; Melvin Johnson; Dmitry Lepikhin; Alexandre Passos; Siamak Shakeri; Emanuel Taropa; Paige Bailey; Zhifeng Chen; Eric Chu; Jonathan H. Clark; Laurent El Shafey; Yanping Huang; Kathy Meier-Hellstern; Gaurav Mishra; Erica Moreira; Mark Omernick; Kevin Robinson; Sebastian Ruder; Yi Tay; Kefan Xiao; Yuanzhong Xu; Yujing Zhang; Gustavo Hernandez Abrego; Junwhan Ahn; Jacob Austin; Paul Barham; Jan Botha; James Bradbury; Siddhartha Brahma; Kevin Brooks; Michele Catasta; Yong Cheng; Colin Cherry; Christopher A. Choquette-Choo; Aakanksha Chowdhery; Clément Crepy; Shachi Dave; Mostafa Dehghani; Sunipa Dev; Jacob Devlin; Mark Díaz; Nan Du; Ethan Dyer; Vlad Feinberg; Fangxiaoyu Feng; Vlad Fienber; Markus Freitag; Xavier Garcia; Sebastian Gehrmann; Lucas Gonzalez; Guy Gur-Ari; Steven Hand; Hadi Hashemi; Le Hou; Joshua Howland; Andrea Hu; Jeffrey Hui; Jeremy Hurwitz; Michael Isard; Abe Ittycheriah; Matthew Jagielski; Wenhao Jia; Kathleen Kenealy; Maxim Krikun; Sneha Kudugunta; Chang Lan; Katherine Lee; Benjamin Lee; Eric Li; Music Li; Wei Li; YaGuang Li; Jian Li; Hyeontaek Lim; Hanzhao Lin; Zhongtao Liu; Frederick Liu; Marcello Maggioni; Aroma Mahendru; Joshua Maynez; Vedant Misra; Maysam Moussalem; Zachary Nado; John Nham; Eric Ni; Andrew Nystrom; Alicia Parrish; Marie Pellat; Martin Polacek; Alex Polozov; Reiner Pope; Siyuan Qiao; Emily Reif; Bryan Richter; Parker Riley; Alex Castro Ros; Aurko Roy; Brennan Saeta; Rajkumar Samuel; Renee Shelby; Ambrose Slone; Daniel Smilkov; David R. So; Daniel Sohn; Simon Tokumine; Dasha Valter; Vijay Vasudevan; Kiran Vodrahalli; Xuezhi Wang; Pidong Wang; Zirui Wang; Tao Wang; John Wieting; Yuhuai Wu; Kelvin Xu; Yunhan Xu; Linting Xue; Pengcheng Yin; Jiahui Yu; Qiao Zhang; Steven Zheng; Ce Zheng; Weikang Zhou; Denny Zhou; Slav Petrov; Yonghui Wu 2023-05-16 Iterative Adversarial Attack on Image-guided Story Ending Generation. (99%)Youze Wang; Wenbo Hu; Richang Hong Releasing Inequality Phenomena in $L_{\infty}$-Adversarial Training via Input Gradient Distillation. (98%)Junxi Chen; Junhao Dong; Xiaohua Xie Ortho-ODE: Enhancing Robustness and of Neural ODEs against Adversarial Attacks. (54%)Vishal Purohit Unlearnable Examples Give a False Sense of Security: Piercing through Unexploitable Data with Learnable Examples. (50%)Wan Jiang; Yunfeng Diao; He Wang; Jianxin Sun; Meng Wang; Richang Hong 2023-05-15 Attacking Perceptual Similarity Metrics. (99%)Abhijay Ghildyal; Feng Liu Exploiting Frequency Spectrum of Adversarial Images for General Robustness. (96%)Chun Yang Tan; Kazuhiko Kawamoto; Hiroshi Kera Training Neural Networks without Backpropagation: A Deeper Dive into the Likelihood Ratio Method. (4%)Jinyang Jiang; Zeliang Zhang; Chenliang Xu; Zhaofei Yu; Yijie Peng Assessing Hidden Risks of LLMs: An Empirical Study on Robustness, Consistency, and Credibility. (1%)Wentao Ye; Mingfeng Ou; Tianyi Li; Yipeng chen; Xuetao Ma; Yifan Yanggong; Sai Wu; Jie Fu; Gang Chen; Haobo Wang; Junbo Zhao 2023-05-14 Diffusion Models for Imperceptible and Transferable Adversarial Attack. (99%)Jianqi Chen; Hao Chen; Keyan Chen; Yilan Zhang; Zhengxia Zou; Zhenwei Shi Improving Defensive Distillation using Teacher Assistant. (96%)Maniratnam Mandal; Suna Gao Manipulating Visually-aware Federated Recommender Systems and Its Countermeasures. (82%)Wei Yuan; Shilong Yuan; Chaoqun Yang; Quoc Viet Hung Nguyen; Hongzhi Yin Watermarking Text Generated by Black-Box Language Models. (9%)Xi Yang; Kejiang Chen; Weiming Zhang; Chang Liu; Yuang Qi; Jie Zhang; Han Fang; Nenghai Yu 2023-05-13 DNN-Defender: An in-DRAM Deep Neural Network Defense Mechanism for Adversarial Weight Attack. (86%)Ranyang Zhou; Sabbir Ahmed; Adnan Siraj Rakin; Shaahin Angizi On enhancing the robustness of Vision Transformers: Defensive Diffusion. (76%)Raza Imam; Muhammad Huzaifa; Mohammed El-Amine Azz Decision-based iterative fragile watermarking for model integrity verification. (50%)Zhaoxia Yin; Heng Yin; Hang Su; Xinpeng Zhang; Zhenzhe Gao 2023-05-12 Efficient Search of Comprehensively Robust Neural Architectures via Multi-fidelity Evaluation. (73%)Jialiang Sun; Wen Yao; Tingsong Jiang; Xiaoqian Chen Adversarial Security and Differential Privacy in mmWave Beam Prediction in 6G networks. (68%)Ghanta Sai Krishna; Kundrapu Supriya; Sanskar Singh; Sabur Baidya Mastering Percolation-like Games with Deep Learning. (1%)Michael M. Danziger; Omkar R. Gojala; Sean P. Cornelius 2023-05-11 Distracting Downpour: Adversarial Weather Attacks for Motion Estimation. (74%)Jenny Schmalfuss; Lukas Mehl; Andrés Bruhn Backdoor Attack with Sparse and Invisible Trigger. (68%)Yinghua Gao; Yiming Li; Xueluan Gong; Shu-Tao Xia; Qian Wang Watch This Space: Securing Satellite Communication through Resilient Transmitter Fingerprinting. (1%)Joshua Smailes; Sebastian Kohler; Simon Birnbach; Martin Strohmeier; Ivan Martinovic 2023-05-10 A Black-Box Attack on Code Models via Representation Nearest Neighbor Search. (99%)Jie Zhang; Wei Ma; Qiang Hu; Shangqing Liu; Xiaofei Xie; Yves Le Traon; Yang Liu Inter-frame Accelerate Attack against Video Interpolation Models. (99%)Junpei Liao; Zhikai Chen; Liang Yi; Wenyuan Yang; Baoyuan Wu; Xiaochun Cao Randomized Smoothing with Masked Inference for Adversarially Robust Text Classifications. (98%)Han Cheol Moon; Shafiq Joty; Ruochen Zhao; Megh Thakkar; Xu Chi Stealthy Low-frequency Backdoor Attack against Deep Neural Networks. (80%)Xinrui Liu; Yu-an Tan; Yajie Wang; Kefan Qiu; Yuanzhang Li Towards Invisible Backdoor Attacks in the Frequency Domain against Deep Neural Networks. (75%)Xinrui Liu; Yajie Wang; Yu-an Tan; Kefan Qiu; Yuanzhang Li The Robustness of Computer Vision Models against Common Corruptions: a Survey. (50%)Shunxin Wang; Raymond Veldhuis; Nicola Strisciuglio An Empirical Study on the Robustness of the Segment Anything Model (SAM). (22%)Yuqing Wang; Yun Zhao; Linda Petzold Robust multi-agent coordination via evolutionary generation of auxiliary adversarial attackers. (12%)Lei Yuan; Zi-Qian Zhang; Ke Xue; Hao Yin; Feng Chen; Cong Guan; Li-He Li; Chao Qian; Yang Yu 2023-05-09 Quantization Aware Attack: Enhancing the Transferability of Adversarial Attacks across Target Models with Different Quantization Bitwidths. (99%)Yulong Yang; Chenhao Lin; Qian Li; Chao Shen; Dawei Zhou; Nannan Wang; Tongliang Liu Attack Named Entity Recognition by Entity Boundary Interference. (98%)Yifei Yang; Hongqiu Wu; Hai Zhao VSMask: Defending Against Voice Synthesis Attack via Real-Time Predictive Perturbation. (96%)Yuanda Wang; Hanqing Guo; Guangjing Wang; Bocheng Chen; Qiben Yan Investigating the Corruption Robustness of Image Classifiers with Random Lp-norm Corruptions. (75%)Georg Siedel; Weijia Shao; Silvia Vock; Andrey Morozov On the Relation between Sharpness-Aware Minimization and Adversarial Robustness. (56%)Zeming Wei; Jingyu Zhu; Yihao Zhang Effects of Real-Life Traffic Sign Alteration on YOLOv7- an Object Recognition Model. (13%)Farhin Farhad Riya; Shahinul Hoque; Md Saif Hassan Onim; Edward Michaud; Edmon Begoli; Jinyuan Stella Sun Turning Privacy-preserving Mechanisms against Federated Learning. (9%)Marco Arazzi; Mauro Conti; Antonino Nocera; Stjepan Picek BadCS: A Backdoor Attack Framework for Code search. (8%)Shiyi Qi; Yuanhang Yang; Shuzhzeng Gao; Cuiyun Gao; Zenglin Xu Quantum Machine Learning for Malware Classification. (1%)Grégoire Barrué; Tony Quertier 2023-05-08 Toward Adversarial Training on Contextualized Language Representation. (93%)Hongqiu Wu; Yongxiang Liu; Hanwen Shi; Hai Zhao; Min Zhang Understanding Noise-Augmented Training for Randomized Smoothing. (64%)Ambar Pal; Jeremias Sulam TAPS: Connecting Certified and Adversarial Training. (41%)Yuhao Mao; Mark Niklas Müller; Marc Fischer; Martin Vechev Privacy-preserving Adversarial Facial Features. (22%)Zhibo Wang; He Wang; Shuaifan Jin; Wenwen Zhang; Jiahui Hu; Yan Wang; Peng Sun; Wei Yuan; Kaixin Liu; Kui Ren Communication-Robust Multi-Agent Learning by Adaptable Auxiliary Multi-Agent Adversary Generation. (1%)Lei Yuan; Feng Chen; Zhongzhang Zhang; Yang Yu 2023-05-07 Adversarial Examples Detection with Enhanced Image Difference Features based on Local Histogram Equalization. (99%)Zhaoxia Yin; Shaowei Zhu; Hang Su; Jianteng Peng; Wanli Lyu; Bin Luo Pick your Poison: Undetectability versus Robustness in Data Poisoning Attacks against Deep Image Classification. (93%)Nils Lukas; Florian Kerschbaum 2023-05-06 The Best Defense is Attack: Repairing Semantics in Textual Adversarial Examples. (99%)Heng Yang; Ke Li Beyond the Model: Data Pre-processing Attack to Deep Learning Models in Android Apps. (92%)Ye Sang; Yujin Huang; Shuo Huang; Helei Cui Towards Prompt-robust Face Privacy Protection via Adversarial Decoupling Augmentation Framework. (38%)Ruijia Wu; Yuhang Wang; Huafeng Shi; Zhipeng Yu; Yichao Wu; Ding Liang Text-to-Image Diffusion Models can be Easily Backdoored through Multimodal Data Poisoning. (2%)Shengfang Zhai; Yinpeng Dong; Qingni Shen; Shi Pu; Yuejian Fang; Hang Su 2023-05-05 White-Box Multi-Objective Adversarial Attack on Dialogue Generation. (99%)Yufei Li; Zexin Li; Yingfan Gao; Cong Liu Evading Watermark based Detection of AI-Generated Content. (87%)Zhengyuan Jiang; Jinghuai Zhang; Neil Zhenqiang Gong Verifiable Learning for Robust Tree Ensembles. (15%)Stefano Calzavara; Lorenzo Cazzaro; Giulio Ermanno Pibiri; Nicola Prezza Repairing Deep Neural Networks Based on Behavior Imitation. (4%)Zhen Liang; Taoran Wu; Changyuan Zhao; Wanwei Liu; Bai Xue; Wenjing Yang; Ji Wang 2023-05-04 Madvex: Instrumentation-based Adversarial Attacks on Machine Learning Malware Detection. (99%)Nils Loose; Felix Mächtle; Claudius Pott; Volodymyr Bezsmertnyi; Thomas Eisenbarth IMAP: Intrinsically Motivated Adversarial Policy. (99%)Xiang Zheng; Xingjun Ma; Shengjie Wang; Xinyu Wang; Chao Shen; Cong Wang Single Node Injection Label Specificity Attack on Graph Neural Networks via Reinforcement Learning. (78%)Dayuan Chen; Jian Zhang; Yuqian Lv; Jinhuan Wang; Hongjie Ni; Shanqing Yu; Zhen Wang; Qi Xuan Faulting original McEliece's implementations is possible: How to mitigate this risk? (2%)Vincent Giraud; Guillaume Bouffard 2023-05-03 New Adversarial Image Detection Based on Sentiment Analysis. (99%)Yulong Wang; Tianxiang Li; Shenghong Li; Xin Yuan; Wei Ni LearnDefend: Learning to Defend against Targeted Model-Poisoning Attacks on Federated Learning. (84%)Kiran Purohit; Soumi Das; Sourangshu Bhattacharya; Santu Rana Defending against Insertion-based Textual Backdoor Attacks via Attribution. (61%)Jiazhao Li; Zhuofeng Wu; Wei Ping; Chaowei Xiao; V. G. Vinod Vydiswaran On the Security Risks of Knowledge Graph Reasoning. (26%)Zhaohan Xi; Tianyu Du; Changjiang Li; Ren Pang; Shouling Ji; Xiapu Luo; Xusheng Xiao; Fenglong Ma; Ting Wang Backdoor Learning on Sequence to Sequence Models. (5%)Lichang Chen; Minhao Cheng; Heng Huang Rethinking Graph Lottery Tickets: Graph Sparsity Matters. (2%)Bo Hui; Da Yan; Xiaolong Ma; Wei-Shinn Ku PTP: Boosting Stability and Performance of Prompt Tuning with Perturbation-Based Regularizer. (1%)Lichang Chen; Heng Huang; Minhao Cheng 2023-05-02 Boosting Adversarial Transferability via Fusing Logits of Top-1 Decomposed Feature. (99%)Juanjuan Weng; Zhiming Luo; Dazhen Lin; Shaozi Li; Zhun Zhong DABS: Data-Agnostic Backdoor attack at the Server in Federated Learning. (73%)Wenqiang Sun; Sen Li; Yuchang Sun; Jun Zhang Towards Imperceptible Document Manipulations against Neural Ranking Models. (67%)Xuanang Chen; Ben He; Zheng Ye; Le Sun; Yingfei Sun Sentiment Perception Adversarial Attacks on Neural Machine Translation Systems. (50%)Vyas Raina; Mark Gales Prompt as Triggers for Backdoor Attack: Examining the Vulnerability in Language Models. (8%)Shuai Zhao; Jinming Wen; Luu Anh Tuan; Junbo Zhao; Jie Fu 2023-05-01 Attack-SAM: Towards Evaluating Adversarial Robustness of Segment Anything Model. (99%)Chenshuang Zhang; Chaoning Zhang; Taegoo Kang; Donghun Kim; Sung-Ho Bae; In So Kweon Physical Adversarial Attacks for Surveillance: A Survey. (98%)Kien Nguyen; Tharindu Fernando; Clinton Fookes; Sridha Sridharan Revisiting Robustness in Graph Machine Learning. (98%)Lukas Gosch; Daniel Sturm; Simon Geisler; Stephan Günnemann Stratified Adversarial Robustness with Rejection. (96%)Jiefeng Chen; Jayaram Raghuram; Jihye Choi; Xi Wu; Yingyu Liang; Somesh Jha Poisoning Language Models During Instruction Tuning. (2%)Alexander Wan; Eric Wallace; Sheng Shen; Dan Klein 2023-04-30 Assessing Vulnerabilities of Adversarial Learning Algorithm through Poisoning Attacks. (98%)Jingfeng Zhang; Bo Song; Bo Han; Lei Liu; Gang Niu; Masashi Sugiyama 2023-04-29 FedGrad: Mitigating Backdoor Attacks in Federated Learning Through Local Ultimate Gradients Inspection. (81%)Thuy Dung Nguyen; Anh Duy Nguyen; Kok-Seng Wong; Huy Hieu Pham; Thanh Hung Nguyen; Phi Le Nguyen; Truong Thao Nguyen Enhancing Adversarial Contrastive Learning via Adversarial Invariant Regularization. (33%)Xilie Xu; Jingfeng Zhang; Feng Liu; Masashi Sugiyama; Mohan Kankanhalli Adversarial Representation Learning for Robust Privacy Preservation in Audio. (1%)Shayan Gharib; Minh Tran; Diep Luong; Konstantinos Drossos; Tuomas Virtanen 2023-04-28 Topic-oriented Adversarial Attacks against Black-box Neural Ranking Models. (99%)Yu-An Liu; Ruqing Zhang; Jiafeng Guo; Rijke Maarten de; Wei Chen; Yixing Fan; Xueqi Cheng On the existence of solutions to adversarial training in multiclass classification. (75%)Nicolas Garcia Trillos; Matt Jacobs; Jakwang Kim The Power of Typed Affine Decision Structures: A Case Study. (3%)Gerrit Nolte; Maximilian Schlüter; Alnis Murtovi; Bernhard Steffen faulTPM: Exposing AMD fTPMs' Deepest Secrets. (3%)Hans Niklas Jacob; Christian Werling; Robert Buhren; Jean-Pierre Seifert SAM Meets Robotic Surgery: An Empirical Study in Robustness Perspective. (1%)An Wang; Mobarakol Islam; Mengya Xu; Yang Zhang; Hongliang Ren 2023-04-27 Adversary Aware Continual Learning. (80%)Muhammad Umer; Robi Polikar Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving. (75%)Zhiyuan Cheng; Hongjun Choi; James Liang; Shiwei Feng; Guanhong Tao; Dongfang Liu; Michael Zuzak; Xiangyu Zhang Boosting Big Brother: Attacking Search Engines with Encodings. (68%)Nicholas Boucher; Luca Pajola; Ilia Shumailov; Ross Anderson; Mauro Conti ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox Generative Model Trigger. (62%)Jiazhao Li; Yijin Yang; Zhuofeng Wu; V. G. Vinod Vydiswaran; Chaowei Xiao Improve Video Representation with Temporal Adversarial Augmentation. (26%)Jinhao Duan; Quanfu Fan; Hao Cheng; Xiaoshuang Shi; Kaidi Xu Origin Tracing and Detecting of LLMs. (1%)Linyang Li; Pengyu Wang; Ke Ren; Tianxiang Sun; Xipeng Qiu Deep Intellectual Property Protection: A Survey. (1%)Yuchen Sun; Tianpeng Liu; Panhe Hu; Qing Liao; Shaojing Fu; Nenghai Yu; Deke Guo; Yongxiang Liu; Li Liu Interactive Greybox Penetration Testing for Cloud Access Control using IAM Modeling and Deep Reinforcement Learning. (1%)Yang Hu; Wenxi Wang; Sarfraz Khurshid; Mohit Tiwari 2023-04-26 Improving Adversarial Transferability via Intermediate-level Perturbation Decay. (98%)Qizhang Li; Yiwen Guo; Wangmeng Zuo; Hao Chen Detection of Adversarial Physical Attacks in Time-Series Image Data. (92%)Ramneet Kaur; Yiannis Kantaros; Wenwen Si; James Weimer; Insup Lee Blockchain-based Federated Learning with SMPC Model Verification Against Poisoning Attack for Healthcare Systems. (13%)Aditya Pribadi Kalapaaking; Ibrahim Khalil; Xun Yi 2023-04-25 Improving Robustness Against Adversarial Attacks with Deeply Quantized Neural Networks. (99%)Ferheen Ayaz; Idris Zakariyya; José Cano; Sye Loong Keoh; Jeremy Singer; Danilo Pau; Mounia Kharbouche-Harrari Generating Adversarial Examples with Task Oriented Multi-Objective Optimization. (99%)Anh Bui; Trung Le; He Zhao; Quan Tran; Paul Montague; Dinh Phung SHIELD: Thwarting Code Authorship Attribution. (98%)Mohammed Abuhamad; Changhun Jung; David Mohaisen; DaeHun Nyang Lyapunov-Stable Deep Equilibrium Models. (82%)Haoyu Chu; Shikui Wei; Ting Liu; Yao Zhao; Yuto Miyatake LSTM-based Load Forecasting Robustness Against Noise Injection Attack in Microgrid. (1%)Amirhossein Nazeri; Pierluigi Pisu 2023-04-24 Evaluating Adversarial Robustness on Document Image Classification. (99%)Timothée Fronteau; Arnaud Paran; Aymen Shabou Combining Adversaries with Anti-adversaries in Training. (64%)Xiaoling Zhou; Nan Yang; Ou Wu Enhancing Fine-Tuning Based Backdoor Defense with Sharpness-Aware Minimization. (41%)Mingli Zhu; Shaokui Wei; Li Shen; Yanbo Fan; Baoyuan Wu Opinion Control under Adversarial Network Perturbation: A Stackelberg Game Approach. (10%)Yuejiang Li; Zhanjiang Chen; H. Vicky Zhao Robust Tickets Can Transfer Better: Drawing More Transferable Subnetworks in Transfer Learning. (1%)Yonggan Fu; Ye Yuan; Shang Wu; Jiayi Yuan; Yingyan Lin 2023-04-23 StyLess: Boosting the Transferability of Adversarial Examples. (99%)Kaisheng Liang; Bin Xiao Evading DeepFake Detectors via Adversarial Statistical Consistency. (98%)Yang Hou; Qing Guo; Yihao Huang; Xiaofei Xie; Lei Ma; Jianjun Zhao 2023-04-22 Detecting Adversarial Faces Using Only Real Face Self-Perturbations. (98%)Qian Wang; Yongqin Xian; Hefei Ling; Jinyuan Zhang; Xiaorui Lin; Ping Li; Jiazhong Chen; Ning Yu Universal Adversarial Backdoor Attacks to Fool Vertical Federated Learning in Cloud-Edge Collaboration. (70%)Peng Chen; Xin Du; Zhihui Lu; Hongfeng Chai 2023-04-21 Launching a Robust Backdoor Attack under Capability Constrained Scenarios. (92%)Ming Yi; Yixiao Xu; Kangyi Ding; Mingyong Yin; Xiaolei Liu Individual Fairness in Bayesian Neural Networks. (69%)Alice Doherty; Matthew Wicker; Luca Laurenti; Andrea Patane Denial-of-Service or Fine-Grained Control: Towards Flexible Model Poisoning Attacks on Federated Learning. (64%)Hangtao Zhang; Zeming Yao; Leo Yu Zhang; Shengshan Hu; Chao Chen; Alan Liew; Zhetao Li Interpretable and Robust AI in EEG Systems: A Survey. (12%)Xinliang Zhou; Chenyu Liu; Liming Zhai; Ziyu Jia; Cuntai Guan; Yang Liu MAWSEO: Adversarial Wiki Search Poisoning for Illicit Online Promotion. (2%)Zilong Lin; Zhengyi Li; Xiaojing Liao; XiaoFeng Wang; Xiaozhong Liu 2023-04-20 Towards the Universal Defense for Query-Based Audio Adversarial Attacks. (99%)Feng Guo; Zheng Sun; Yuxuan Chen; Lei Ju Diversifying the High-level Features for better Adversarial Transferability. (99%)Zhiyuan Wang; Zeliang Zhang; Siyuan Liang; Xiaosen Wang Using Z3 for Formal Modeling and Verification of FNN Global Robustness. (98%)Yihao Zhang; Zeming Wei; Xiyue Zhang; Meng Sun Certified Adversarial Robustness Within Multiple Perturbation Bounds. (96%)Soumalya Nandi; Sravanti Addepalli; Harsh Rangwani; R. Venkatesh Babu Can Perturbations Help Reduce Investment Risks? Risk-Aware Stock Recommendation via Split Variational Adversarial Training. (93%)Jiezhu Cheng; Kaizhu Huang; Zibin Zheng Adversarial Infrared Blocks: A Black-box Attack to Thermal Infrared Detectors at Multiple Angles in Physical World. (89%)Chengyin Hu; Weiwen Shi; Tingsong Jiang; Wen Yao; Ling Tian; Xiaoqian Chen An Analysis of the Completion Time of the BB84 Protocol. (22%)Sounak Kar; Jean-Yves Le Boudec A Plug-and-Play Defensive Perturbation for Copyright Protection of DNN-based Applications. (13%)Donghua Wang; Wen Yao; Tingsong Jiang; Weien Zhou; Lang Lin; Xiaoqian Chen Enhancing object detection robustness: A synthetic and natural perturbation approach. (12%)Nilantha Premakumara; Brian Jalaian; Niranjan Suri; Hooman Samani RoCOCO: Robustness Benchmark of MS-COCO to Stress-test Image-Text Matching Models. (8%)Seulki Park; Daeho Um; Hajung Yoon; Sanghyuk Chun; Sangdoo Yun; Jin Young Choi Get Rid Of Your Trail: Remotely Erasing Backdoors in Federated Learning. (2%)Manaar Alam; Hithem Lamri; Michail Maniatakos Learning Sample Difficulty from Pre-trained Models for Reliable Prediction. (1%)Peng Cui; Dan Zhang; Zhijie Deng; Yinpeng Dong; Jun Zhu 2023-04-19 Jedi: Entropy-based Localization and Removal of Adversarial Patches. (84%)Bilel Tarchoun; Anouar Ben Khalifa; Mohamed Ali Mahjoub; Nael Abu-Ghazaleh; Ihsen Alouani GREAT Score: Global Robustness Evaluation of Adversarial Perturbation using Generative Models. (81%)Zaitang Li; Pin-Yu Chen; Tsung-Yi Ho Secure Split Learning against Property Inference, Data Reconstruction, and Feature Space Hijacking Attacks. (5%)Yunlong Mao; Zexi Xin; Zhenyu Li; Jue Hong; Qingyou Yang; Sheng Zhong Density-Insensitive Unsupervised Domain Adaption on 3D Object Detection. (1%)Qianjiang Hu; Daizong Liu; Wei Hu On the Robustness of Aspect-based Sentiment Analysis: Rethinking Model, Data, and Training. (1%)Hao Fei; Tat-Seng Chua; Chenliang Li; Donghong Ji; Meishan Zhang; Yafeng Ren Fundamental Limitations of Alignment in Large Language Models. (1%)Yotam Wolf; Noam Wies; Oshri Avnery; Yoav Levine; Amnon Shashua 2023-04-18 Wavelets Beat Monkeys at Adversarial Robustness. (99%)Jingtong Su; Julia Kempe Towards the Transferable Audio Adversarial Attack via Ensemble Methods. (99%)Feng Guo; Zheng Sun; Yuxuan Chen; Lei Ju Masked Language Model Based Textual Adversarial Example Detection. (99%)Xiaomei Zhang; Zhaoxi Zhang; Qi Zhong; Xufei Zheng; Yanjun Zhang; Shengshan Hu; Leo Yu Zhang In ChatGPT We Trust? Measuring and Characterizing the Reliability of ChatGPT. (80%)Xinyue Shen; Zeyuan Chen; Michael Backes; Yang Zhang Generative models improve fairness of medical classifiers under distribution shifts. (13%)Ira Ktena; Olivia Wiles; Isabela Albuquerque; Sylvestre-Alvise Rebuffi; Ryutaro Tanno; Abhijit Guha Roy; Shekoofeh Azizi; Danielle Belgrave; Pushmeet Kohli; Alan Karthikesalingam; Taylan Cemgil; Sven Gowal 2023-04-17 Evil from Within: Machine Learning Backdoors through Hardware Trojans. (15%)Alexander Warnecke; Julian Speith; Jan-Niklas Möller; Konrad Rieck; Christof Paar GrOVe: Ownership Verification of Graph Neural Networks using Embeddings. (13%)Asim Waheed; Vasisht Duddu; N. Asokan OOD-CV-v2: An extended Benchmark for Robustness to Out-of-Distribution Shifts of Individual Nuisances in Natural Images. (1%)Bingchen Zhao; Jiahao Wang; Wufei Ma; Artur Jesslen; Siwei Yang; Shaozuo Yu; Oliver Zendel; Christian Theobalt; Alan Yuille; Adam Kortylewski 2023-04-16 A Random-patch based Defense Strategy Against Physical Attacks for Face Recognition Systems. (98%)JiaHao Xie; Ye Luo; Jianwei Lu RNN-Guard: Certified Robustness Against Multi-frame Attacks for Recurrent Neural Networks. (96%)Yunruo Zhang; Tianyu Du; Shouling Ji; Peng Tang; Shanqing Guo JoB-VS: Joint Brain-Vessel Segmentation in TOF-MRA Images. (15%)Natalia Valderrama; Ioannis Pitsiorlas; Luisa Vargas; Pablo Arbeláez; Maria A. Zuluaga 2023-04-14 Interpretability is a Kind of Safety: An Interpreter-based Ensemble for Adversary Defense. (99%)Jingyuan Wang; Yufan Wu; Mingxuan Li; Xin Lin; Junjie Wu; Chao Li Combining Generators of Adversarial Malware Examples to Increase Evasion Rate. (99%)Matouš Kozák; Martin Jureček Cross-Entropy Loss Functions: Theoretical Analysis and Applications. (3%)Anqi Mao; Mehryar Mohri; Yutao Zhong Pool Inference Attacks on Local Differential Privacy: Quantifying the Privacy Guarantees of Apple's Count Mean Sketch in Practice. (2%)Andrea Gadotti; Florimond Houssiau; Meenatchi Sundaram Muthu Selva Annamalai; Montjoye Yves-Alexandre de 2023-04-13 Generating Adversarial Examples with Better Transferability via Masking Unimportant Parameters of Surrogate Model. (99%)Dingcheng Yang; Wenjian Yu; Zihao Xiao; Jiaqi Luo Certified Zeroth-order Black-Box Defense with Robust UNet Denoiser. (96%)Astha Verma; Siddhesh Bangar; A V Subramanyam; Naman Lal; Rajiv Ratn Shah; Shin'ichi Satoh False Claims against Model Ownership Resolution. (93%)Jian Liu; Rui Zhang; Sebastian Szyller; Kui Ren; N. Asokan Adversarial Examples from Dimensional Invariance. (45%)Benjamin L. Badger Understanding Overfitting in Adversarial Training in Kernel Regression. (1%)Teng Zhang; Kang Li LSFSL: Leveraging Shape Information in Few-shot Learning. (1%)Deepan Chakravarthi Padmanabhan; Shruthi Gowda; Elahe Arani; Bahram Zonooz 2023-04-12 Generative Adversarial Networks-Driven Cyber Threat Intelligence Detection Framework for Securing Internet of Things. (92%)Mohamed Amine Ferrag; Djallel Hamouda; Merouane Debbah; Leandros Maglaras; Abderrahmane Lakas Exploiting Logic Locking for a Neural Trojan Attack on Machine Learning Accelerators. (1%)Hongye Xu; Dongfang Liu; Cory Merkel; Michael Zuzak 2023-04-11 RecUP-FL: Reconciling Utility and Privacy in Federated Learning via User-configurable Privacy Defense. (99%)Yue Cui; Syed Irfan Ali Meerza; Zhuohang Li; Luyang Liu; Jiaxin Zhang; Jian Liu Simultaneous Adversarial Attacks On Multiple Face Recognition System Components. (98%)Inderjeet Singh; Kazuya Kakizaki; Toshinori Araki Boosting Cross-task Transferability of Adversarial Patches with Visual Relations. (98%)Tony Ma; Songze Li; Yisong Xiao; Shunchang Liu Benchmarking the Physical-world Adversarial Robustness of Vehicle Detection. (92%)Tianyuan Zhang; Yisong Xiao; Xiaoya Zhang; Hao Li; Lu Wang On the Adversarial Inversion of Deep Biometric Representations. (67%)Gioacchino Tangari; Shreesh Keskar; Hassan Jameel Asghar; Dali Kaafar Overload: Latency Attacks on Object Detection for Edge Devices. (33%)Erh-Chung Chen; Pin-Yu Chen; I-Hsin Chung; Che-rung Lee Towards More Robust and Accurate Sequential Recommendation with Cascade-guided Adversarial Training. (9%)Juntao Tan; Shelby Heinecke; Zhiwei Liu; Yongjun Chen; Yongfeng Zhang; Huan Wang 2023-04-10 Generating Adversarial Attacks in the Latent Space. (98%)Nitish Shukla; Sudipta Banerjee Reinforcement Learning-Based Black-Box Model Inversion Attacks. (67%)Gyojin Han; Jaehyun Choi; Haeil Lee; Junmo Kim Defense-Prefix for Preventing Typographic Attacks on CLIP. (16%)Hiroki Azuma; Yusuke Matsui Helix++: A platform for efficiently securing software. (1%)Jack W. Davidson; Jason D. Hiser; Anh Nguyen-Tuong 2023-04-09 Certifiable Black-Box Attack: Ensuring Provably Successful Attack for Adversarial Examples. (99%)Hanbin Hong; Yuan Hong Adversarially Robust Neural Architecture Search for Graph Neural Networks. (80%)Beini Xie; Heng Chang; Ziwei Zhang; Xin Wang; Daixin Wang; Zhiqiang Zhang; Rex Ying; Wenwu Zhu Unsupervised Multi-Criteria Adversarial Detection in Deep Image Retrieval. (68%)Yanru Xiao; Cong Wang; Xing Gao 2023-04-08 Robust Deep Learning Models Against Semantic-Preserving Adversarial Attack. (99%)Dashan Gao; Yunce Zhao; Yinghua Yao; Zeqi Zhang; Bifei Mao; Xin Yao RobCaps: Evaluating the Robustness of Capsule Networks against Affine Transformations and Adversarial Attacks. (98%)Alberto Marchisio; Marco Antonio De; Alessio Colucci; Maurizio Martina; Muhammad Shafique Exploring the Connection between Robust and Generative Models. (67%)Senad Beadini; Iacopo Masi Benchmarking the Robustness of Quantized Models. (47%)Yisong Xiao; Tianyuan Zhang; Shunchang Liu; Haotong Qin Attack is Good Augmentation: Towards Skeleton-Contrastive Representation Learning. (13%)Binqian Xu; Xiangbo Shu; Rui Yan; Guo-Sen Xie; Yixiao Ge; Mike Zheng Shou Deep Prototypical-Parts Ease Morphological Kidney Stone Identification and are Competitively Robust to Photometric Perturbations. (4%)Daniel Flores-Araiza; Francisco Lopez-Tiro; Jonathan El-Beze; Jacques Hubert; Miguel Gonzalez-Mendoza; Gilberto Ochoa-Ruiz; Christian Daul EMP-SSL: Towards Self-Supervised Learning in One Training Epoch. (1%)Shengbang Tong; Yubei Chen; Yi Ma; Yann Lecun 2023-04-07 Architecture-Preserving Provable Repair of Deep Neural Networks. (1%)Zhe Tao; Stephanie Nawas; Jacqueline Mitchell; Aditya V. Thakur ASPEST: Bridging the Gap Between Active Learning and Selective Prediction. (1%)Jiefeng Chen; Jinsung Yoon; Sayna Ebrahimi; Sercan Arik; Somesh Jha; Tomas Pfister 2023-04-06 Quantifying and Defending against Privacy Threats on Federated Knowledge Graph Embedding. (45%)Yuke Hu; Wei Liang; Ruofan Wu; Kai Xiao; Weiqiang Wang; Xiaochen Li; Jinfei Liu; Zhan Qin Manipulating Federated Recommender Systems: Poisoning with Synthetic Users and Its Countermeasures. (45%)Wei Yuan; Quoc Viet Hung Nguyen; Tieke He; Liang Chen; Hongzhi Yin Improving Visual Question Answering Models through Robustness Analysis and In-Context Learning with a Chain of Basic Questions. (10%)Jia-Hong Huang; Modar Alfadly; Bernard Ghanem; Marcel Worring EZClone: Improving DNN Model Extraction Attack via Shape Distillation from GPU Execution Profiles. (4%)Jonah O'Brien Weiss; Tiago Alves; Sandip Kundu Evaluating the Robustness of Machine Reading Comprehension Models to Low Resource Entity Renaming. (2%)Clemencia Siro; Tunde Oluwaseyi Ajayi Rethinking Evaluation Protocols of Visual Representations Learned via Self-supervised Learning. (1%)Jae-Hun Lee; Doyoung Yoon; ByeongMoon Ji; Kyungyul Kim; Sangheum Hwang Reliable Learning for Test-time Attacks and Distribution Shift. (1%)Maria-Florina Balcan; Steve Hanneke; Rattana Pukdee; Dravyansh Sharma Benchmarking Robustness to Text-Guided Corruptions. (1%)Mohammadreza Mofayezi; Yasamin Medghalchi 2023-04-05 A Certified Radius-Guided Attack Framework to Image Segmentation Models. (99%)Wenjie Qu; Youqi Li; Binghui Wang How to choose your best allies for a transferable attack? (99%)Thibault Maho; Seyed-Mohsen Moosavi-Dezfooli; Teddy Furon Going Further: Flatness at the Rescue of Early Stopping for Adversarial Example Transferability. (99%)Martin Gubri; Maxime Cordy; Yves Le Traon Robust Neural Architecture Search. (92%)Xunyu Zhu; Jian Li; Yong Liu; Weiping Wang Hyper-parameter Tuning for Adversarially Robust Models. (62%)Pedro Mendes; Paolo Romano; David Garlan JPEG Compressed Images Can Bypass Protections Against AI Editing. (15%)Pedro Sandoval-Segura; Jonas Geiping; Tom Goldstein FACE-AUDITOR: Data Auditing in Facial Recognition Systems. (1%)Min Chen; Zhikun Zhang; Tianhao Wang; Michael Backes; Yang Zhang 2023-04-04 CGDTest: A Constrained Gradient Descent Algorithm for Testing Neural Networks. (31%)Vineel Nagisetty; Laura Graves; Guanting Pan; Piyush Jha; Vijay Ganesh Selective Knowledge Sharing for Privacy-Preserving Federated Distillation without A Good Teacher. (1%)Jiawei Shao; Fangzhao Wu; Jun Zhang EGC: Image Generation and Classification via a Single Energy-Based Model. (1%)Qiushan Guo; Chuofan Ma; Yi Jiang; Zehuan Yuan; Yizhou Yu; Ping Luo 2023-04-03 Defending Against Patch-based Backdoor Attacks on Self-Supervised Learning. (76%)Ajinkya Tejankar; Maziar Sanjabi; Qifan Wang; Sinong Wang; Hamed Firooz; Hamed Pirsiavash; Liang Tan Model-Agnostic Reachability Analysis on Deep Neural Networks. (75%)Chi Zhang; Wenjie Ruan; Fu Wang; Peipei Xu; Geyong Min; Xiaowei Huang NetFlick: Adversarial Flickering Attacks on Deep Learning Based Video Compression. (69%)Jung-Woo Chang; Nojan Sheybani; Shehzeen Samarah Hussain; Mojan Javaheripi; Seira Hidano; Farinaz Koushanfar Learning About Simulated Adversaries from Human Defenders using Interactive Cyber-Defense Games. (1%)Baptiste Prebot; Yinuo Du; Cleotilde Gonzalez 2023-04-01 GradMDM: Adversarial Attack on Dynamic Networks. (84%)Jianhong Pan; Lin Geng Foo; Qichen Zheng; Zhipeng Fan; Hossein Rahmani; Qiuhong Ke; Jun Liu Instance-level Trojan Attacks on Visual Question Answering via Adversarial Learning in Neuron Activation Space. (67%)Yuwei Sun; Hideya Ochiai; Jun Sakuma 2023-03-31 Improving Fast Adversarial Training with Prior-Guided Knowledge. (99%)Xiaojun Jia; Yong Zhang; Xingxing Wei; Baoyuan Wu; Ke Ma; Jue Wang; Xiaochun Sr Cao To be Robust and to be Fair: Aligning Fairness with Robustness. (93%)Junyi Chai; Xiaoqian Wang Fooling Polarization-based Vision using Locally Controllable Polarizing Projection. (91%)Zhuoxiao Li; Zhihang Zhong; Shohei Nobuhara; Ko Nishino; Yinqiang Zheng Per-Example Gradient Regularization Improves Learning Signals from Noisy Data. (3%)Xuran Meng; Yuan Cao; Difan Zou Secure Federated Learning against Model Poisoning Attacks via Client Filtering. (2%)Duygu Nur Yaldiz; Tuo Zhang; Salman Avestimehr DIME-FM: DIstilling Multimodal and Efficient Foundation Models. (1%)Ximeng Sun; Pengchuan Zhang; Peizhao Zhang; Hardik Shah; Kate Saenko; Xide Xia A Generative Framework for Low-Cost Result Validation of Outsourced Machine Learning Tasks. (1%)Abhinav Kumar; Miguel A. Guirao Aguilera; Reza Tourani; Satyajayant Misra 2023-03-30 Adversarial Attack and Defense for Dehazing Networks. (97%)Jie Gui; Xiaofeng Cong; Chengwei Peng; Yuan Yan Tang; James Tin-Yau Kwok Generating Adversarial Samples in Mini-Batches May Be Detrimental To Adversarial Robustness. (96%)Timothy Redgrave; Colton Crum Towards Adversarially Robust Continual Learning. (95%)Tao Bai; Chen Chen; Lingjuan Lyu; Jun Zhao; Bihan Wen Understanding the Robustness of 3D Object Detection with Bird's-Eye-View Representations in Autonomous Driving. (81%)Zijian Zhu; Yichi Zhang; Hai Chen; Yinpeng Dong; Shu Zhao; Wenbo Ding; Jiachen Zhong; Shibao Zheng Robo3D: Towards Robust and Reliable 3D Perception against Corruptions. (2%)Lingdong Kong; Youquan Liu; Xin Li; Runnan Chen; Wenwei Zhang; Jiawei Ren; Liang Pan; Kai Chen; Ziwei Liu Establishing baselines and introducing TernaryMixOE for fine-grained out-of-distribution detection. (1%)Noah Fleischmann; Walter Bennette; Nathan Inkawhich Explainable Intrusion Detection Systems Using Competitive Learning Techniques. (1%)Jesse Ables; Thomas Kirby; Sudip Mittal; Ioana Banicescu; Shahram Rahimi; William Anderson; Maria Seale Differential Area Analysis for Ransomware: Attacks, Countermeasures, and Limitations. (1%)Marco Venturini; Francesco Freda; Emanuele Miotto; Alberto Giaretta; Mauro Conti 2023-03-29 Latent Feature Relation Consistency for Adversarial Robustness. (99%)Xingbin Liu; Huafeng Kuang; Hong Liu; Xianming Lin; Yongjian Wu; Rongrong Ji Beyond Empirical Risk Minimization: Local Structure Preserving Regularization for Improving Adversarial Robustness. (99%)Wei Wei; Jiahuan Zhou; Ying Wu Targeted Adversarial Attacks on Wind Power Forecasts. (88%)René Heinrich; Christoph Scholz; Stephan Vogt; Malte Lehna Towards Reasonable Budget Allocation in Untargeted Graph Structure Attacks via Gradient Debias. (67%)Zihan Liu; Yun Luo; Lirong Wu; Zicheng Liu; Stan Z. Li ImageNet-E: Benchmarking Neural Network Robustness via Attribute Editing. (56%)Xiaodan Li; Yuefeng Chen; Yao Zhu; Shuhui Wang; Rong Zhang; Hui Xue Graph Neural Networks for Hardware Vulnerability Analysis -- Can you Trust your GNN? (16%)Lilas Alrahis; Ozgur Sinanoglu Mole Recruitment: Poisoning of Image Classifiers via Selective Batch Sampling. (10%)Ethan Wisdom; Tejas Gokhale; Chaowei Xiao; Yezhou Yang A Tensor-based Convolutional Neural Network for Small Dataset Classification. (2%)Zhenhua Chen; David Crandall ALUM: Adversarial Data Uncertainty Modeling from Latent Model Uncertainty Compensation. (1%)Wei Wei; Jiahuan Zhou; Hongze Li; Ying Wu 2023-03-28 A Pilot Study of Query-Free Adversarial Attack against Stable Diffusion. (99%)Haomin Zhuang; Yihua Zhang; Sijia Liu Improving the Transferability of Adversarial Samples by Path-Augmented Method. (99%)Jianping Zhang; Jen-tse Huang; Wenxuan Wang; Yichen Li; Weibin Wu; Xiaosen Wang; Yuxin Su; Michael R. Lyu Towards Effective Adversarial Textured 3D Meshes on Physical Face Recognition. (99%)Xiao Yang; Chang Liu; Longlong Xu; Yikai Wang; Yinpeng Dong; Ning Chen; Hang Su; Jun Zhu Transferable Adversarial Attacks on Vision Transformers with Token Gradient Regularization. (98%)Jianping Zhang; Yizhan Huang; Weibin Wu; Michael R. Lyu Denoising Autoencoder-based Defensive Distillation as an Adversarial Robustness Algorithm. (98%)Bakary Badjie; José Cecílio; António Casimiro TransAudio: Towards the Transferable Adversarial Audio Attack via Learning Contextualized Perturbations. (98%)Qi Gege; Yuefeng Chen; Xiaofeng Mao; Yao Zhu; Binyuan Hui; Xiaodan Li; Rong Zhang; Hui Xue A Survey on Malware Detection with Graph Representation Learning. (41%)Tristan Bilot; Nour El Madhoun; Khaldoun Al Agha; Anis Zouaoui Provable Robustness for Streaming Models with a Sliding Window. (15%)Aounon Kumar; Vinu Sankar Sadasivan; Soheil Feizi Machine-learned Adversarial Attacks against Fault Prediction Systems in Smart Electrical Grids. (9%)Carmelo Ardito; Yashar Deldjoo; Noia Tommaso Di; Sciascio Eugenio Di; Fatemeh Nazary; Giovanni Servedio On the Use of Reinforcement Learning for Attacking and Defending Load Frequency Control. (3%)Amr S. Mohamed; Deepa Kundur A Universal Identity Backdoor Attack against Speaker Verification based on Siamese Network. (1%)Haodong Zhao; Wei Du; Junjie Guo; Gongshen Liu 2023-03-27 Classifier Robustness Enhancement Via Test-Time Transformation. (99%)Tsachi Blau; Roy Ganz; Chaim Baskin; Michael Elad; Alex Bronstein Improving the Transferability of Adversarial Examples via Direction Tuning. (99%)Xiangyuan Yang; Jie Lin; Hanlin Zhang; Xinyu Yang; Peng Zhao EMShepherd: Detecting Adversarial Samples via Side-channel Leakage. (99%)Ruyi Ding; Cheng Gongye; Siyue Wang; Aidong Ding; Yunsi Fei Learning the Unlearnable: Adversarial Augmentations Suppress Unlearnable Example Attacks. (97%)Tianrui Qin; Xitong Gao; Juanjuan Zhao; Kejiang Ye; Cheng-Zhong Xu Detecting Backdoors During the Inference Stage Based on Corruption Robustness Consistency. (76%)Xiaogeng Liu; Minghui Li; Haoyu Wang; Shengshan Hu; Dengpan Ye; Hai Jin; Libing Wu; Chaowei Xiao CAT:Collaborative Adversarial Training. (69%)Xingbin Liu; Huafeng Kuang; Xianming Lin; Yongjian Wu; Rongrong Ji Diffusion Denoised Smoothing for Certified and Adversarial Robust Out-Of-Distribution Detection. (67%)Nicola Franco; Daniel Korth; Jeanette Miriam Lorenz; Karsten Roscher; Stephan Guennemann Personalized Federated Learning on Long-Tailed Data via Adversarial Feature Augmentation. (41%)Yang Lu; Pinxin Qian; Gang Huang; Hanzi Wang Mask and Restore: Blind Backdoor Defense at Test Time with Masked Autoencoder. (41%)Tao Sun; Lu Pang; Chao Chen; Haibin Ling Sequential training of GANs against GAN-classifiers reveals correlated "knowledge gaps" present among independently trained GAN instances. (41%)Arkanath Pathak; Nicholas Dufour Anti-DreamBooth: Protecting users from personalized text-to-image synthesis. (5%)Le Thanh Van; Hao Phung; Thuan Hoang Nguyen; Quan Dao; Ngoc Tran; Anh Tran 2023-03-26 MGTBench: Benchmarking Machine-Generated Text Detection. (61%)Xinlei He; Xinyue Shen; Zeyuan Chen; Michael Backes; Yang Zhang 2023-03-25 AdvCheck: Characterizing Adversarial Examples via Local Gradient Checking. (99%)Ruoxi Chen; Haibo Jin; Jinyin Chen; Haibin Zheng CFA: Class-wise Calibrated Fair Adversarial Training. (98%)Zeming Wei; Yifei Wang; Yiwen Guo; Yisen Wang PORE: Provably Robust Recommender Systems against Data Poisoning Attacks. (68%)Jinyuan Jia; Yupei Liu; Yuepeng Hu; Neil Zhenqiang Gong Improving robustness of jet tagging algorithms with adversarial training: exploring the loss surface. (12%)Annika Stein 2023-03-24 PIAT: Parameter Interpolation based Adversarial Training for Image Classification. (99%)Kun He; Xin Liu; Yichen Yang; Zhou Qin; Weigao Wen; Hui Xue; John E. Hopcroft How many dimensions are required to find an adversarial example? (99%)Charles Godfrey; Henry Kvinge; Elise Bishoff; Myles Mckay; Davis Brown; Tim Doster; Eleanor Byler Effective black box adversarial attack with handcrafted kernels. (99%)Petr Dvořáček; Petr Hurtik; Petra Števuliáková Adversarial Attack and Defense for Medical Image Analysis: Methods and Applications. (99%)Junhao Dong; Junxi Chen; Xiaohua Xie; Jianhuang Lai; Hao Chen Improved Adversarial Training Through Adaptive Instance-wise Loss Smoothing. (99%)Lin Li; Michael Spratling Feature Separation and Recalibration for Adversarial Robustness. (98%)Woo Jae Kim; Yoonki Cho; Junsik Jung; Sung-Eui Yoon Physically Adversarial Infrared Patches with Learnable Shapes and Locations. (97%)Wei Xingxing; Yu Jie; Huang Yao Generalist: Decoupling Natural and Robust Generalization. (96%)Hongjun Wang; Yisen Wang Ensemble-based Blackbox Attacks on Dense Prediction. (86%)Zikui Cai; Yaoteng Tan; M. Salman Asif Backdoor Attacks with Input-unique Triggers in NLP. (54%)Xukun Zhou; Jiwei Li; Tianwei Zhang; Lingjuan Lyu; Muqiao Yang; Jun He PoisonedGNN: Backdoor Attack on Graph Neural Networks-based Hardware Security Systems. (22%)Lilas Alrahis; Satwik Patnaik; Muhammad Abdullah Hanif; Muhammad Shafique; Ozgur Sinanoglu Enhancing Multiple Reliability Measures via Nuisance-extended Information Bottleneck. (5%)Jongheon Jeong; Sihyun Yu; Hankook Lee; Jinwoo Shin Optimal Smoothing Distribution Exploration for Backdoor Neutralization in Deep Learning-based Traffic Systems. (2%)Yue Wang; Wending Li; Michail Maniatakos; Saif Eddin Jabari TRAK: Attributing Model Behavior at Scale. (1%)Sung Min Park; Kristian Georgiev; Andrew Ilyas; Guillaume Leclerc; Aleksander Madry 2023-03-23 Watch Out for the Confusing Faces: Detecting Face Swapping with the Probability Distribution of Face Identification Models. (68%)Yuxuan Duan; Xuhong Zhang; Chuer Yu; Zonghui Wang; Shouling Ji; Wenzhi Chen Quadratic Graph Attention Network (Q-GAT) for Robust Construction of Gene Regulatory Networks. (50%)Hui Zhang; Xuexin An; Qiang He; Yudong Yao; Feng-Lei Fan; Yueyang Teng Optimization and Optimizers for Adversarial Robustness. (41%)Hengyue Liang; Buyun Liang; Le Peng; Ying Cui; Tim Mitchell; Ju Sun Adversarial Robustness and Feature Impact Analysis for Driver Drowsiness Detection. (41%)João Vitorino; Lourenço Rodrigues; Eva Maia; Isabel Praça; André Lourenço Paraphrasing evades detectors of AI-generated text, but retrieval is an effective defense. (15%)Kalpesh Krishna; Yixiao Song; Marzena Karpinska; John Wieting; Mohit Iyyer Decentralized Adversarial Training over Graphs. (13%)Ying Cao; Elsa Rizk; Stefan Vlaski; Ali H. Sayed Don't FREAK Out: A Frequency-Inspired Approach to Detecting Backdoor Poisoned Samples in DNNs. (8%)Hasan Abed Al Kader Hammoud; Adel Bibi; Philip H. S. Torr; Bernard Ghanem Low-frequency Image Deep Steganography: Manipulate the Frequency Distribution to Hide Secrets with Tenacious Robustness. (1%)Huajie Chen; Tianqing Zhu; Yuan Zhao; Bo Liu; Xin Yu; Wanlei Zhou Efficient Symbolic Reasoning for Neural-Network Verification. (1%)Zi Dj Wang; Somesh Dj Jha; Dj Krishnamurthy; Dvijotham 2023-03-22 Reliable and Efficient Evaluation of Adversarial Robustness for Deep Hashing-Based Retrieval. (99%)Xunguang Wang; Jiawang Bai; Xinyue Xu; Xiaomeng Li Semantic Image Attack for Visual Model Diagnosis. (99%)Jinqi Luo; Zhaoning Wang; Chen Henry Wu; Dong Huang; la Torre Fernando De Revisiting DeepFool: generalization and improvement. (99%)Alireza Abdollahpourrostam; Mahed Abroshan; Seyed-Mohsen Moosavi-Dezfooli Wasserstein Adversarial Examples on Univariant Time Series Data. (99%)Wenjie Wang; Li Xiong; Jian Lou Test-time Defense against Adversarial Attacks: Detection and Reconstruction of Adversarial Examples via Masked Autoencoder. (99%)Yun-Yun Tsai; Ju-Chin Chao; Albert Wen; Zhaoyuan Yang; Chengzhi Mao; Tapan Shah; Junfeng Yang Sibling-Attack: Rethinking Transferable Adversarial Attacks against Face Recognition. (78%)Zexin Li; Bangjie Yin; Taiping Yao; Juefeng Guo; Shouhong Ding; Simin Chen; Cong Liu An Extended Study of Human-like Behavior under Adversarial Training. (76%)Paul Gavrikov; Janis Keuper; Margret Keuper Distribution-restrained Softmax Loss for the Model Robustness. (38%)Hao Wang; Chen Li; Jinzhe Jiang; Xin Zhang; Yaqian Zhao; Weifeng Gong Backdoor Defense via Adaptively Splitting Poisoned Dataset. (16%)Kuofeng Gao; Yang Bai; Jindong Gu; Yong Yang; Shu-Tao Xia Edge Deep Learning Model Protection via Neuron Authorization. (11%)Jinyin Chen; Haibin Zheng; Tao Liu; Rongchang Li; Yao Cheng; Xuhong Zhang; Shouling Ji 2023-03-21 State-of-the-art optical-based physical adversarial attacks for deep learning computer vision systems. (99%)Junbin Fang; You Jiang; Canjian Jiang; Zoe L. Jiang; Siu-Ming Yiu; Chuanyi Liu Information-containing Adversarial Perturbation for Combating Facial Manipulation Systems. (99%)Yao Zhu; Yuefeng Chen; Xiaodan Li; Rong Zhang; Xiang Tian; Bolun Zheng; Yaowu Chen Bridging Optimal Transport and Jacobian Regularization by Optimal Trajectory for Enhanced Adversarial Defense. (99%)Binh M. Le; Shahroz Tariq; Simon S. Woo Efficient Decision-based Black-box Patch Attacks on Video Recognition. (98%)Kaixun Jiang; Zhaoyu Chen; Hao Huang; Jiafeng Wang; Dingkang Yang; Bo Li; Yan Wang; Wenqiang Zhang Black-box Backdoor Defense via Zero-shot Image Purification. (86%)Yucheng Shi; Mengnan Du; Xuansheng Wu; Zihan Guan; Jin Sun; Ninghao Liu Out of Thin Air: Exploring Data-Free Adversarial Robustness Distillation. (10%)Yuzheng Wang; Zhaoyu Chen; Dingkang Yang; Pinxue Guo; Kaixun Jiang; Wenqiang Zhang; Lizhe Qi Influencer Backdoor Attack on Semantic Segmentation. (10%)Haoheng Lan; Jindong Gu; Philip Torr; Hengshuang Zhao LOKI: Large-scale Data Reconstruction Attack against Federated Learning through Model Manipulation. (9%)Joshua C. Zhao; Atul Sharma; Ahmed Roushdy Elkordy; Yahya H. Ezzeldin; Salman Avestimehr; Saurabh Bagchi Poisoning Attacks in Federated Edge Learning for Digital Twin 6G-enabled IoTs: An Anticipatory Study. (1%)Mohamed Amine Ferrag; Burak Kantarci; Lucas C. Cordeiro; Merouane Debbah; Kim-Kwang Raymond Choo 2023-03-20 TWINS: A Fine-Tuning Framework for Improved Transferability of Adversarial Robustness and Generalization. (99%)Ziquan Liu; Yi Xu; Xiangyang Ji; Antoni B. Chan Adversarial Attacks against Binary Similarity Systems. (99%)Gianluca Capozzi; Daniele Cono D'Elia; Luna Giuseppe Antonio Di; Leonardo Querzoni DRSM: De-Randomized Smoothing on Malware Classifier Providing Certified Robustness. (99%)Shoumik Saha; Wenxiao Wang; Yigitcan Kaya; Soheil Feizi; Tudor Dumitras Translate your gibberish: black-box adversarial attack on machine translation systems. (83%)Andrei Chertkov; Olga Tsymboi; Mikhail Pautov; Ivan Oseledets GNN-Ensemble: Towards Random Decision Graph Neural Networks. (56%)Wenqi Wei; Mu Qiao; Divyesh Jadav Benchmarking Robustness of 3D Object Detection to Common Corruptions in Autonomous Driving. (41%)Yinpeng Dong; Caixin Kang; Jinlai Zhang; Zijian Zhu; Yikai Wang; Xiao Yang; Hang Su; Xingxing Wei; Jun Zhu Did You Train on My Dataset? Towards Public Dataset Protection with Clean-Label Backdoor Watermarking. (9%)Ruixiang Tang; Qizhang Feng; Ninghao Liu; Fan Yang; Xia Hu Boosting Semi-Supervised Learning by Exploiting All Unlabeled Data. (2%)Yuhao Chen; Xin Tan; Borui Zhao; Zhaowei Chen; Renjie Song; Jiajun Liang; Xuequan Lu Make Landscape Flatter in Differentially Private Federated Learning. (1%)Yifan Shi; Yingqi Liu; Kang Wei; Li Shen; Xueqian Wang; Dacheng Tao Robustifying Token Attention for Vision Transformers. (1%)Yong Guo; David Stutz; Bernt Schiele 2023-03-19 Randomized Adversarial Training via Taylor Expansion. (99%)Gaojie Jin; Xinping Yi; Dengyu Wu; Ronghui Mu; Xiaowei Huang AdaptGuard: Defending Against Universal Attacks for Model Adaptation. (82%)Lijun Sheng; Jian Liang; Ran He; Zilei Wang; Tieniu Tan 2023-03-18 NoisyHate: Benchmarking Content Moderation Machine Learning Models with Human-Written Perturbations Online. (98%)Yiran Ye; Thai Le; Dongwon Lee FedRight: An Effective Model Copyright Protection for Federated Learning. (96%)Jinyin Chen; Mingjun Li; Mingjun Li; Haibin Zheng 2023-03-17 Fuzziness-tuned: Improving the Transferability of Adversarial Examples. (99%)Xiangyuan Yang; Jie Lin; Hanlin Zhang; Xinyu Yang; Peng Zhao It Is All About Data: A Survey on the Effects of Data on Adversarial Robustness. (99%)Peiyu Xiong; Michael Tegegn; Jaskeerat Singh Sarin; Shubhraneel Pal; Julia Rubin Robust Mode Connectivity-Oriented Adversarial Defense: Enhancing Neural Network Robustness Against Diversified $\ell_p$ Attacks. (99%)Ren Wang; Yuxuan Li; Sijia Liu Detection of Uncertainty in Exceedance of Threshold (DUET): An Adversarial Patch Localizer. (83%)Terence Jie Chua; Wenhan Yu; Jun Zhao Can AI-Generated Text be Reliably Detected? (45%)Vinu Sankar Sadasivan; Aounon Kumar; Sriram Balasubramanian; Wenxiao Wang; Soheil Feizi Adversarial Counterfactual Visual Explanations. (31%)Guillaume Jeanneret; Loïc Simon; Frédéric Jurie MedLocker: A Transferable Adversarial Watermarking for Preventing Unauthorized Analysis of Medical Image Dataset. (16%)Bangzheng Pu; Xingxing Wei; Shiji Zhao; Huazhu Fu Mobile Edge Adversarial Detection for Digital Twinning to the Metaverse with Deep Reinforcement Learning. (9%)Terence Jie Chua; Wenhan Yu; Jun Zhao Moving Target Defense for Service-oriented Mission-critical Networks. (1%)Doğanalp Ergenç; Florian Schneider; Peter Kling; Mathias Fischer 2023-03-16 Rethinking Model Ensemble in Transfer-based Adversarial Attacks. (99%)Huanran Chen; Yichi Zhang; Yinpeng Dong; Jun Zhu Class Attribute Inference Attacks: Inferring Sensitive Class Information by Diffusion-Based Attribute Manipulations. (68%)Lukas Struppek; Dominik Hintersdorf; Felix Friedrich; Manuel Brack; Patrick Schramowski; Kristian Kersting Among Us: Adversarially Robust Collaborative Perception by Consensus. (67%)Yiming Li; Qi Fang; Jiamu Bai; Siheng Chen; Felix Juefei-Xu; Chen Feng Exorcising ''Wraith'': Protecting LiDAR-based Object Detector in Automated Driving System from Appearing Attacks. (50%)Qifan Xiao; Xudong Pan; Yifan Lu; Mi Zhang; Jiarun Dai; Min Yang Rethinking White-Box Watermarks on Deep Learning Models under Neural Structural Obfuscation. (11%)Yifan Yan; Xudong Pan; Mi Zhang; Min Yang 2023-03-15 Black-box Adversarial Example Attack towards FCG Based Android Malware Detection under Incomplete Feature Information. (99%)Heng Li; Zhang Cheng; Bang Wu; Liheng Yuan; Cuiying Gao; Wei Yuan; Xiapu Luo Robust Evaluation of Diffusion-Based Adversarial Purification. (83%)Minjong Lee; Dongwoo Kim DeeBBAA: A benchmark Deep Black Box Adversarial Attack against Cyber-Physical Power Systems. (81%)Arnab Bhattacharjee; Tapan K. Saha; Ashu Verma; Sukumar Mishra The Devil's Advocate: Shattering the Illusion of Unexploitable Data using Diffusion Models. (67%)Hadi M. Dolatabadi; Sarah Erfani; Christopher Leckie EvalAttAI: A Holistic Approach to Evaluating Attribution Maps in Robust and Non-Robust Models. (45%)Ian E. Nielsen; Ravi P. Ramachandran; Nidhal Bouaynaya; Hassan M. Fathallah-Shaykh; Ghulam Rasool Agnostic Multi-Robust Learning Using ERM. (12%)Saba Ahmadi; Avrim Blum; Omar Montasser; Kevin Stangl Reinforce Data, Multiply Impact: Improved Model Accuracy and Robustness with Dataset Reinforcement. (1%)Fartash Faghri; Hadi Pouransari; Sachin Mehta; Mehrdad Farajtabar; Ali Farhadi; Mohammad Rastegari; Oncel Tuzel GPT-4 Technical Report. (1%)Rai OpenAI; Josh Rai Achiam; Steven Rai Adler; Sandhini Rai Agarwal; Lama Rai Ahmad; Ilge Rai Akkaya; Florencia Leoni Rai Aleman; Diogo Rai Almeida; Janko Rai Altenschmidt; Sam Rai Altman; Shyamal Rai Anadkat; Red Rai Avila; Igor Rai Babuschkin; Suchir Rai Balaji; Valerie Rai Balcom; Paul Rai Baltescu; Haiming Rai Bao; Mohammad Rai Bavarian; Jeff Rai Belgum; Irwan Rai Bello; Jake Rai Berdine; Gabriel Rai Bernadett-Shapiro; Christopher Rai Berner; Lenny Rai Bogdonoff; Oleg Rai Boiko; Madelaine Rai Boyd; Anna-Luisa Rai Brakman; Greg Rai Brockman; Tim Rai Brooks; Miles Rai Brundage; Kevin Rai Button; Trevor Rai Cai; Rosie Rai Campbell; Andrew Rai Cann; Brittany Rai Carey; Chelsea Rai Carlson; Rory Rai Carmichael; Brooke Rai Chan; Che Rai Chang; Fotis Rai Chantzis; Derek Rai Chen; Sully Rai Chen; Ruby Rai Chen; Jason Rai Chen; Mark Rai Chen; Ben Rai Chess; Chester Rai Cho; Casey Rai Chu; Hyung Won Rai Chung; Dave Rai Cummings; Jeremiah Rai Currier; Yunxing Rai Dai; Cory Rai Decareaux; Thomas Rai Degry; Noah Rai Deutsch; Damien Rai Deville; Arka Rai Dhar; David Rai Dohan; Steve Rai Dowling; Sheila Rai Dunning; Adrien Rai Ecoffet; Atty Rai Eleti; Tyna Rai Eloundou; David Rai Farhi; Liam Rai Fedus; Niko Rai Felix; Simón Posada Rai Fishman; Juston Rai Forte; Isabella Rai Fulford; Leo Rai Gao; Elie Rai Georges; Christian Rai Gibson; Vik Rai Goel; Tarun Rai Gogineni; Gabriel Rai Goh; Rapha Rai Gontijo-Lopes; Jonathan Rai Gordon; Morgan Rai Grafstein; Scott Rai Gray; Ryan Rai Greene; Joshua Rai Gross; Shixiang Shane Rai Gu; Yufei Rai Guo; Chris Rai Hallacy; Jesse Rai Han; Jeff Rai Harris; Yuchen Rai He; Mike Rai Heaton; Johannes Rai Heidecke; Chris Rai Hesse; Alan Rai Hickey; Wade Rai Hickey; Peter Rai Hoeschele; Brandon Rai Houghton; Kenny Rai Hsu; Shengli Rai Hu; Xin Rai Hu; Joost Rai Huizinga; Shantanu Rai Jain; Shawn Rai Jain; Joanne Rai Jang; Angela Rai Jiang; Roger Rai Jiang; Haozhun Rai Jin; Denny Rai Jin; Shino Rai Jomoto; Billie Rai Jonn; Heewoo Rai Jun; Tomer Rai Kaftan; Łukasz Rai Kaiser; Ali Rai Kamali; Ingmar Rai Kanitscheider; Nitish Shirish Rai Keskar; Tabarak Rai Khan; Logan Rai Kilpatrick; Jong Wook Rai Kim; Christina Rai Kim; Yongjik Rai Kim; Jan Hendrik Rai Kirchner; Jamie Rai Kiros; Matt Rai Knight; Daniel Rai Kokotajlo; Łukasz Rai Kondraciuk; Andrew Rai Kondrich; Aris Rai Konstantinidis; Kyle Rai Kosic; Gretchen Rai Krueger; Vishal Rai Kuo; Michael Rai Lampe; Ikai Rai Lan; Teddy Rai Lee; Jan Rai Leike; Jade Rai Leung; Daniel Rai Levy; Chak Ming Rai Li; Rachel Rai Lim; Molly Rai Lin; Stephanie Rai Lin; Mateusz Rai Litwin; Theresa Rai Lopez; Ryan Rai Lowe; Patricia Rai Lue; Anna Rai Makanju; Kim Rai Malfacini; Sam Rai Manning; Todor Rai Markov; Yaniv Rai Markovski; Bianca Rai Martin; Katie Rai Mayer; Andrew Rai Mayne; Bob Rai McGrew; Scott Mayer Rai McKinney; Christine Rai McLeavey; Paul Rai McMillan; Jake Rai McNeil; David Rai Medina; Aalok Rai Mehta; Jacob Rai Menick; Luke Rai Metz; Andrey Rai Mishchenko; Pamela Rai Mishkin; Vinnie Rai Monaco; Evan Rai Morikawa; Daniel Rai Mossing; Tong Rai Mu; Mira Rai Murati; Oleg Rai Murk; David Rai Mély; Ashvin Rai Nair; Reiichiro Rai Nakano; Rajeev Rai Nayak; Arvind Rai Neelakantan; Richard Rai Ngo; Hyeonwoo Rai Noh; Long Rai Ouyang; Cullen Rai O'Keefe; Jakub Rai Pachocki; Alex Rai Paino; Joe Rai Palermo; Ashley Rai Pantuliano; Giambattista Rai Parascandolo; Joel Rai Parish; Emy Rai Parparita; Alex Rai Passos; Mikhail Rai Pavlov; Andrew Rai Peng; Adam Rai Perelman; Filipe de Avila Belbute Rai Peres; Michael Rai Petrov; Henrique Ponde de Oliveira Rai Pinto; Rai Michael; Pokorny; Michelle Pokrass; Vitchyr H. Pong; Tolly Powell; Alethea Power; Boris Power; Elizabeth Proehl; Raul Puri; Alec Radford; Jack Rae; Aditya Ramesh; Cameron Raymond; Francis Real; Kendra Rimbach; Carl Ross; Bob Rotsted; Henri Roussez; Nick Ryder; Mario Saltarelli; Ted Sanders; Shibani Santurkar; Girish Sastry; Heather Schmidt; David Schnurr; John Schulman; Daniel Selsam; Kyla Sheppard; Toki Sherbakov; Jessica Shieh; Sarah Shoker; Pranav Shyam; Szymon Sidor; Eric Sigler; Maddie Simens; Jordan Sitkin; Katarina Slama; Ian Sohl; Benjamin Sokolowsky; Yang Song; Natalie Staudacher; Felipe Petroski Such; Natalie Summers; Ilya Sutskever; Jie Tang; Nikolas Tezak; Madeleine B. Thompson; Phil Tillet; Amin Tootoonchian; Elizabeth Tseng; Preston Tuggle; Nick Turley; Jerry Tworek; Juan Felipe Cerón Uribe; Andrea Vallone; Arun Vijayvergiya; Chelsea Voss; Carroll Wainwright; Justin Jay Wang; Alvin Wang; Ben Wang; Jonathan Ward; Jason Wei; CJ Weinmann; Akila Welihinda; Peter Welinder; Jiayi Weng; Lilian Weng; Matt Wiethoff; Dave Willner; Clemens Winter; Samuel Wolrich; Hannah Wong; Lauren Workman; Sherwin Wu; Jeff Wu; Michael Wu; Kai Xiao; Tao Xu; Sarah Yoo; Kevin Yu; Qiming Yuan; Wojciech Zaremba; Rowan Zellers; Chong Zhang; Marvin Zhang; Shengjia Zhao; Tianhao Zheng; Juntang Zhuang; William Zhuk; Barret Zoph 2023-03-14 Verifying the Robustness of Automatic Credibility Assessment. (99%)Piotr Przybyła; Alexander Shvets; Horacio Saggion Resilient Dynamic Average Consensus based on Trusted agents. (69%)Shamik Bhattacharyya; Rachel Kalpana Kalaimani Improving Adversarial Robustness with Hypersphere Embedding and Angular-based Regularizations. (31%)Olukorede Fakorede; Ashutosh Nirala; Modeste Atsague; Jin Tian 2023-03-13 Constrained Adversarial Learning and its applicability to Automated Software Testing: a systematic review. (99%)João Vitorino; Tiago Dias; Tiago Fonseca; Eva Maia; Isabel Praça Can Adversarial Examples Be Parsed to Reveal Victim Model Information? (99%)Yuguang Yao; Jiancheng Liu; Yifan Gong; Xiaoming Liu; Yanzhi Wang; Xue Lin; Sijia Liu Review on the Feasibility of Adversarial Evasion Attacks and Defenses for Network Intrusion Detection Systems. (99%)Islam Debicha; Benjamin Cochez; Tayeb Kenaza; Thibault Debatty; Jean-Michel Dricot; Wim Mees SMUG: Towards robust MRI reconstruction by smoothed unrolling. (98%)Hui Li; Jinghan Jia; Shijun Liang; Yuguang Yao; Saiprasad Ravishankar; Sijia Liu Model-tuning Via Prompts Makes NLP Models Adversarially Robust. (96%)Mrigank Raman; Pratyush Maini; J. Zico Kolter; Zachary C. Lipton; Danish Pruthi Robust Contrastive Language-Image Pretraining against Adversarial Attacks. (83%)Wenhan Yang; Baharan Mirzasoleiman Model Extraction Attacks on Split Federated Learning. (47%)Jingtao Li; Adnan Siraj Rakin; Xing Chen; Li Yang; Zhezhi He; Deliang Fan; Chaitali Chakrabarti WDiscOOD: Out-of-Distribution Detection via Whitened Linear Discriminative Analysis. (1%)Yiye Chen; Yunzhi Lin; Ruinian Xu; Patricio A. Vela Pixel-wise Gradient Uncertainty for Convolutional Neural Networks applied to Out-of-Distribution Segmentation. (1%)Kira Maag; Tobias Riedlinger 2023-03-12 Adv-Bot: Realistic Adversarial Botnet Attacks against Network Intrusion Detection Systems. (99%)Islam Debicha; Benjamin Cochez; Tayeb Kenaza; Thibault Debatty; Jean-Michel Dricot; Wim Mees Adaptive Local Adversarial Attacks on 3D Point Clouds for Augmented Reality. (99%)Weiquan Liu; Shijun Zheng; Cheng Wang DNN-Alias: Deep Neural Network Protection Against Side-Channel Attacks via Layer Balancing. (96%)Mahya Morid Ahmadi; Lilas Alrahis; Ozgur Sinanoglu; Muhammad Shafique Multi-metrics adaptively identifies backdoors in Federated learning. (92%)Siquan Huang; Yijiang Li; Chong Chen; Leyu Shi; Ying Gao Adversarial Attacks to Direct Data-driven Control for Destabilization. (91%)Hampei Sasahara Backdoor Defense via Deconfounded Representation Learning. (83%)Zaixi Zhang; Qi Liu; Zhicai Wang; Zepu Lu; Qingyong Hu Interpreting Hidden Semantics in the Intermediate Layers of 3D Point Cloud Classification Neural Network. (76%)Weiquan Liu; Minghao Liu; Shijun Zheng; Cheng Wang Boosting Source Code Learning with Data Augmentation: An Empirical Study. (11%)Zeming Dong; Qiang Hu; Yuejun Guo; Zhenya Zhang; Maxime Cordy; Mike Papadakis; Yves Le Traon; Jianjun Zhao 2023-03-11 Improving the Robustness of Deep Convolutional Neural Networks Through Feature Learning. (99%)Jin Ding; Jie-Chao Zhao; Yong-Zhi Sun; Ping Tan; Ji-En Ma; You-Tong Fang SHIELD: An Adaptive and Lightweight Defense against the Remote Power Side-Channel Attacks on Multi-tenant FPGAs. (8%)Mahya Morid Ahmadi; Faiq Khalid; Radha Vaidya; Florian Kriebel; Andreas Steininger; Muhammad Shafique 2023-03-10 Turning Strengths into Weaknesses: A Certified Robustness Inspired Attack Framework against Graph Neural Networks. (99%)Binghui Wang; Meng Pang; Yun Dong Boosting Adversarial Attacks by Leveraging Decision Boundary Information. (99%)Boheng Zeng; LianLi Gao; QiLong Zhang; ChaoQun Li; JingKuan Song; ShuaiQi Jing Adversarial Attacks and Defenses in Machine Learning-Powered Networks: A Contemporary Survey. (99%)Yulong Wang; Tong Sun; Shenghong Li; Xin Yuan; Wei Ni; Ekram Hossain; H. Vincent Poor Investigating Stateful Defenses Against Black-Box Adversarial Examples. (99%)Ryan Feng; Ashish Hooda; Neal Mangaokar; Kassem Fawaz; Somesh Jha; Atul Prakash MIXPGD: Hybrid Adversarial Training for Speech Recognition Systems. (99%)Aminul Huq; Weiyi Zhang; Xiaolin Hu Do we need entire training data for adversarial training? (99%)Vipul Gupta; Apurva Narayan TrojDiff: Trojan Attacks on Diffusion Models with Diverse Targets. (61%)Weixin Chen; Dawn Song; Bo Li Adapting Contrastive Language-Image Pretrained (CLIP) Models for Out-of-Distribution Detection. (13%)Nikolas Adaloglou; Felix Michels; Tim Kaiser; Markus Kollmann 2023-03-09 NoiseCAM: Explainable AI for the Boundary Between Noise and Adversarial Attacks. (99%)Wenkai Tan; Justus Renkhoff; Alvaro Velasquez; Ziyu Wang; Lusi Li; Jian Wang; Shuteng Niu; Fan Yang; Yongxin Liu; Houbing Song Evaluating the Robustness of Conversational Recommender Systems by Adversarial Examples. (92%)Ali Montazeralghaem; James Allan Identification of Systematic Errors of Image Classifiers on Rare Subgroups. (83%)Jan Hendrik Metzen; Robin Hutmacher; N. Grace Hua; Valentyn Boreiko; Dan Zhang Learning the Legibility of Visual Text Perturbations. (78%)Dev Seth; Rickard Stureborg; Danish Pruthi; Bhuwan Dhingra Efficient Certified Training and Robustness Verification of Neural ODEs. (75%)Mustafa Zeqiri; Mark Niklas Müller; Marc Fischer; Martin Vechev Feature Unlearning for Pre-trained GANs and VAEs. (68%)Saemi Moon; Seunghyuk Cho; Dongwoo Kim 2023-03-08 Immune Defense: A Novel Adversarial Defense Mechanism for Preventing the Generation of Adversarial Examples. (99%)Jinwei Wang; Hao Wu; Haihua Wang; Jiawei Zhang; Xiangyang Luo; Bin Ma Decision-BADGE: Decision-based Adversarial Batch Attack with Directional Gradient Estimation. (99%)Geunhyeok Yu; Minwoo Jeon; Hyoseok Hwang Exploring Adversarial Attacks on Neural Networks: An Explainable Approach. (99%)Justus Renkhoff; Wenkai Tan; Alvaro Velasquez; illiam Yichen Wang; Yongxin Liu; Jian Wang; Shuteng Niu; Lejla Begic Fazlic; Guido Dartmann; Houbing Song BeamAttack: Generating High-quality Textual Adversarial Examples through Beam Search and Mixed Semantic Spaces. (99%)Hai Zhu; Qingyang Zhao; Yuren Wu DeepGD: A Multi-Objective Black-Box Test Selection Approach for Deep Neural Networks. (3%)Zohreh Aghababaeyan; Manel Abdellatif; Mahboubeh Dadkhah; Lionel Briand 2023-03-07 Logit Margin Matters: Improving Transferable Targeted Adversarial Attack by Logit Calibration. (99%)Juanjuan Weng; Zhiming Luo; Zhun Zhong; Shaozi Li; Nicu Sebe Patch of Invisibility: Naturalistic Black-Box Adversarial Attacks on Object Detectors. (98%)Raz Lapid; Moshe Sipper Robustness-preserving Lifelong Learning via Dataset Condensation. (96%)Jinghan Jia; Yihua Zhang; Dogyoon Song; Sijia Liu; Alfred Hero CUDA: Convolution-based Unlearnable Datasets. (82%)Vinu Sankar Sadasivan; Mahdi Soltanolkotabi; Soheil Feizi EavesDroid: Eavesdropping User Behaviors via OS Side-Channels on Smartphones. (11%)Quancheng Wang; Ming Tang; Jianming Fu Stabilized training of joint energy-based models and their practical applications. (2%)Martin Sustek; Samik Sadhu; Lukas Burget; Hynek Hermansky; Jesus Villalba; Laureano Moro-Velazquez; Najim Dehak 2023-03-06 CleanCLIP: Mitigating Data Poisoning Attacks in Multimodal Contrastive Learning. (41%)Hritik Bansal; Nishad Singhi; Yu Yang; Fan Yin; Aditya Grover; Kai-Wei Chang Students Parrot Their Teachers: Membership Inference on Model Distillation. (31%)Matthew Jagielski; Milad Nasr; Christopher Choquette-Choo; Katherine Lee; Nicholas Carlini On the Feasibility of Specialized Ability Extracting for Large Language Code Models. (22%)Zongjie Li; Chaozheng Wang; Pingchuan Ma; Chaowei Liu; Shuai Wang; Daoyuan Wu; Cuiyun Gao A Unified Algebraic Perspective on Lipschitz Neural Networks. (15%)Alexandre Araujo; Aaron Havens; Blaise Delattre; Alexandre Allauzen; Bin Hu Learning to Backdoor Federated Learning. (15%)Henger Li; Chen Wu; Sencun Zhu; Zizhan Zheng Partial-Information, Longitudinal Cyber Attacks on LiDAR in Autonomous Vehicles. (10%)R. Spencer Hallyburton; Qingzhao Zhang; Z. Morley Mao; Miroslav Pajic ALMOST: Adversarial Learning to Mitigate Oracle-less ML Attacks via Synthesis Tuning. (1%)Animesh Basak Chowdhury; Lilas Alrahis; Luca Collini; Johann Knechtel; Ramesh Karri; Siddharth Garg; Ozgur Sinanoglu; Benjamin Tan Rethinking Confidence Calibration for Failure Prediction. (1%)Fei Zhu; Zhen Cheng; Xu-Yao Zhang; Cheng-Lin Liu 2023-03-05 Visual Analytics of Neuron Vulnerability to Adversarial Attacks on Convolutional Neural Networks. (99%)Yiran Li; Junpeng Wang; Takanori Fujiwara; Kwan-Liu Ma Consistent Valid Physically-Realizable Adversarial Attack against Crowd-flow Prediction Models. (99%)Hassan Ali; Muhammad Atif Butt; Fethi Filali; Ala Al-Fuqaha; Junaid Qadir Adversarial Sampling for Fairness Testing in Deep Neural Network. (98%)Tosin Ige; William Marfo; Justin Tonkinson; Sikiru Adewale; Bolanle Hafiz Matti Local Environment Poisoning Attacks on Federated Reinforcement Learning. (12%)Evelyn Ma; Rasoul Etesami Robustness, Evaluation and Adaptation of Machine Learning Models in the Wild. (10%)Vihari Piratla Knowledge-Based Counterfactual Queries for Visual Question Answering. (3%)Theodoti Stoikou; Maria Lymperaiou; Giorgos Stamou 2023-03-04 Improved Robustness Against Adaptive Attacks With Ensembles and Error-Correcting Output Codes. (68%)Thomas Philippon; Christian Gagné 2023-03-03 PointCert: Point Cloud Classification with Deterministic Certified Robustness Guarantees. (91%)Jinghuai Zhang; Jinyuan Jia; Hongbin Liu; Neil Zhenqiang Gong Certified Robust Neural Networks: Generalization and Corruption Resistance. (82%)Amine Bennouna; Ryan Lucas; Parys Bart Van AdvART: Adversarial Art for Camouflaged Object Detection Attacks. (75%)Amira Guesmi; Ioan Marius Bilasco; Muhammad Shafique; Ihsen Alouani Backdoor Attacks and Defenses in Federated Learning: Survey, Challenges and Future Research Directions. (47%)Thuy Dung Nguyen; Tuan Nguyen; Phi Le Nguyen; Hieu H. Pham; Khoa Doan; Kok-Seng Wong Adversarial Attacks on Machine Learning in Embedded and IoT Platforms. (38%)Christian Westbrook; Sudeep Pasricha Revisiting Adversarial Training for ImageNet: Architectures, Training and Generalization across Threat Models. (33%)Naman D Singh; Francesco Croce; Matthias Hein Stealthy Perception-based Attacks on Unmanned Aerial Vehicles. (16%)Amir Khazraei; Haocheng Meng; Miroslav Pajic TrojText: Test-time Invisible Textual Trojan Insertion. (2%)Qian Lou; Yepeng Liu; Bo Feng 2023-03-02 Defending against Adversarial Audio via Diffusion Model. (99%)Shutong Wu; Jiongxiao Wang; Wei Ping; Weili Nie; Chaowei Xiao Demystifying Causal Features on Adversarial Examples and Causal Inoculation for Robust Network by Adversarial Instrumental Variable Regression. (99%)Junho Kim. Byung-Kwan Lee; Yong Man Ro AdvRain: Adversarial Raindrops to Attack Camera-based Smart Vision Systems. (99%)Amira Guesmi; Muhammad Abdullah Hanif; Muhammad Shafique APARATE: Adaptive Adversarial Patch for CNN-based Monocular Depth Estimation for Autonomous Navigation. (99%)Amira Guesmi; Muhammad Abdullah Hanif; Ihsen Alouani; Muhammad Shafique Targeted Adversarial Attacks against Neural Machine Translation. (98%)Sahar Sadrizadeh; AmirHossein Dabiri Aghdam; Ljiljana Dolamic; Pascal Frossard The Double-Edged Sword of Implicit Bias: Generalization vs. Robustness in ReLU Networks. (93%)Spencer Frei; Gal Vardi; Peter L. Bartlett; Nathan Srebro Feature Perturbation Augmentation for Reliable Evaluation of Importance Estimators in Neural Networks. (10%)Lennart Brocki; Neo Christopher Chung D-Score: An Expert-Based Method for Assessing the Detectability of IoT-Related Cyber-Attacks. (3%)Yair Meidan; Daniel Benatar; Ron Bitton; Dan Avraham; Asaf Shabtai Interpretable System Identification and Long-term Prediction on Time-Series Data. (1%)Xiaoyi Liu; Duxin Chen; Wenjia Wei; Xia Zhu; Wenwu Yu Consistency Models. (1%)Yang Song; Prafulla Dhariwal; Mark Chen; Ilya Sutskever CADeSH: Collaborative Anomaly Detection for Smart Homes. (1%)Yair Meidan; Dan Avraham; Hanan Libhaber; Asaf Shabtai Conflict-Based Cross-View Consistency for Semi-Supervised Semantic Segmentation. (1%)Zicheng Wang; Zhen Zhao; Xiaoxia Xing; Dong Xu; Xiangyu Kong; Luping Zhou 2023-03-01 To Make Yourself Invisible with Adversarial Semantic Contours. (99%)Yichi Zhang; Zijian Zhu; Hang Su; Jun Zhu; Shibao Zheng; Yuan He; Hui Xue Adversarial Examples Exist in Two-Layer ReLU Networks for Low Dimensional Data Manifolds. (98%)Odelia Melamed; Gilad Yehudai; Gal Vardi Frauds Bargain Attack: Generating Adversarial Text Samples via Word Manipulation Process. (97%)Mingze Ni; Zhensu Sun; Wei Liu A Practical Upper Bound for the Worst-Case Attribution Deviations. (70%)Fan Wang; Adams Wai-Kin Kong Combating Exacerbated Heterogeneity for Robust Models in Federated Learning. (54%)Jianing Zhu; Jiangchao Yao; Tongliang Liu; Quanming Yao; Jianliang Xu; Bo Han Poster: Sponge ML Model Attacks of Mobile Apps. (8%)Souvik Paul; Nicolas Kourtellis DOLOS: A Novel Architecture for Moving Target Defense. (8%)Giulio Pagnotta; Gaspari Fabio De; Dorjan Hitaj; Mauro Andreolini; Michele Colajanni; Luigi V. Mancini Mitigating Backdoors in Federated Learning with FLD. (2%)Yihang Lin; Pengyuan Zhou; Zhiqian Wu; Yong Liao Competence-Based Analysis of Language Models. (1%)Adam Davies; Jize Jiang; ChengXiang Zhai 2023-02-28 A semantic backdoor attack against Graph Convolutional Networks. (98%)Jiazhu Dai; Zhipeng Xiong Single Image Backdoor Inversion via Robust Smoothed Classifiers. (88%)Mingjie Sun; J. Zico Kolter Feature Extraction Matters More: Universal Deepfake Disruption through Attacking Ensemble Feature Extractors. (67%)Long Tang; Dengpan Ye; Zhenhao Lu; Yunming Zhang; Shengshan Hu; Yue Xu; Chuanxi Chen Backdoor Attacks Against Deep Image Compression via Adaptive Frequency Trigger. (11%)Yi Yu; Yufei Wang; Wenhan Yang; Shijian Lu; Yap-peng Tan; Alex C. Kot FreeEagle: Detecting Complex Neural Trojans in Data-Free Cases. (1%)Chong Fu; Xuhong Zhang; Shouling Ji; Ting Wang; Peng Lin; Yanghe Feng; Jianwei Yin 2023-02-27 A Comprehensive Study on Robustness of Image Classification Models: Benchmarking and Rethinking. (99%)Chang Liu; Yinpeng Dong; Wenzhao Xiang; Xiao Yang; Hang Su; Jun Zhu; Yuefeng Chen; Yuan He; Hui Xue; Shibao Zheng Adversarial Attack with Raindrops. (99%)Jiyuan Liu; Bingyi Lu; Mingkang Xiong; Tao Zhang; Huilin Xiong Physical Adversarial Attacks on Deep Neural Networks for Traffic Sign Recognition: A Feasibility Study. (99%)Fabian Woitschek; Georg Schneider Aegis: Mitigating Targeted Bit-flip Attacks against Deep Neural Networks. (98%)Jialai Wang; Ziyuan Zhang; Meiqi Wang; Han Qiu; Tianwei Zhang; Qi Li; Zongpeng Li; Tao Wei; Chao Zhang CBA: Contextual Background Attack against Optical Aerial Detection in the Physical World. (98%)Jiawei Lian; Xiaofei Wang; Yuru Su; Mingyang Ma; Shaohui Mei Improving Model Generalization by On-manifold Adversarial Augmentation in the Frequency Domain. (96%)Chang Liu; Wenzhao Xiang; Yuan He; Hui Xue; Shibao Zheng; Hang Su Efficient and Low Overhead Website Fingerprinting Attacks and Defenses based on TCP/IP Traffic. (83%)Guodong Huang; Chuan Ma; Ming Ding; Yuwen Qian; Chunpeng Ge; Liming Fang; Zhe Liu GLOW: Global Layout Aware Attacks on Object Detection. (81%)Buyu Liu; BaoJun; Jianping Fan; Xi Peng; Kui Ren; Jun Yu Online Black-Box Confidence Estimation of Deep Neural Networks. (16%)Fabian Woitschek; Georg Schneider Implicit Poisoning Attacks in Two-Agent Reinforcement Learning: Adversarial Policies for Training-Time Attacks. (15%)Mohammad Mohammadi; Jonathan Nöther; Debmalya Mandal; Adish Singla; Goran Radanovic Differentially Private Diffusion Models Generate Useful Synthetic Images. (10%)Sahra Ghalebikesabi; Leonard Berrada; Sven Gowal; Ira Ktena; Robert Stanforth; Jamie Hayes; Soham De; Samuel L. Smith; Olivia Wiles; Borja Balle Learning to Retain while Acquiring: Combating Distribution-Shift in Adversarial Data-Free Knowledge Distillation. (5%)Gaurav Patel; Konda Reddy Mopuri; Qiang Qiu 2023-02-26 Contextual adversarial attack against aerial detection in the physical world. (99%)Jiawei Lian; Xiaofei Wang; Yuru Su; Mingyang Ma; Shaohui Mei Randomness in ML Defenses Helps Persistent Attackers and Hinders Evaluators. (96%)Keane Lucas; Matthew Jagielski; Florian Tramèr; Lujo Bauer; Nicholas Carlini 2023-02-25 Deep Learning-based Multi-Organ CT Segmentation with Adversarial Data Augmentation. (99%)Shaoyan Pan; Shao-Yuan Lo; Min Huang; Chaoqiong Ma; Jacob Wynne; Tonghe Wang; Tian Liu; Xiaofeng Yang Scalable Attribution of Adversarial Attacks via Multi-Task Learning. (99%)Zhongyi Guo; Keji Han; Yao Ge; Wei Ji; Yun Li SATBA: An Invisible Backdoor Attack Based On Spatial Attention. (74%)Huasong Zhou; Xiaowei Xu; Xiaodong Wang; Leon Bevan Bullock Bayesian Neural Networks Avoid Encoding Complex and Perturbation-Sensitive Concepts. (1%)Qihan Ren; Huiqi Deng; Yunuo Chen; Siyu Lou; Quanshi Zhang 2023-02-24 Defending Against Backdoor Attacks by Layer-wise Feature Analysis. (68%)Najeeb Moharram Jebreel; Josep Domingo-Ferrer; Yiming Li Chaotic Variational Auto encoder-based Adversarial Machine Learning. (54%)Pavan Venkata Sainadh Reddy; Yelleti Vivek; Gopi Pranay; Vadlamani Ravi Robust Weight Signatures: Gaining Robustness as Easy as Patching Weights? (12%)Ruisi Cai; Zhenyu Zhang; Zhangyang Wang 2023-02-23 Less is More: Data Pruning for Faster Adversarial Training. (99%)Yize Li; Pu Zhao; Xue Lin; Bhavya Kailkhura; Ryan Goldhahn A Plot is Worth a Thousand Words: Model Information Stealing Attacks via Scientific Plots. (99%)Boyang Zhang; Xinlei He; Yun Shen; Tianhao Wang; Yang Zhang Boosting Adversarial Transferability using Dynamic Cues. (99%)Muzammal Naseer; Ahmad Mahmood; Salman Khan; Fahad Khan HyperAttack: Multi-Gradient-Guided White-box Adversarial Structure Attack of Hypergraph Neural Networks. (98%)Chao Hu; Ruishi Yu; Binqi Zeng; Yu Zhan; Ying Fu; Quan Zhang; Rongkai Liu; Heyuan Shi Investigating Catastrophic Overfitting in Fast Adversarial Training: A Self-fitting Perspective. (84%)Zhengbao He; Tao Li; Sizhe Chen; Xiaolin Huang More than you've asked for: A Comprehensive Analysis of Novel Prompt Injection Threats to Application-Integrated Large Language Models. (70%)Kai Greshake; Sahar Abdelnabi; Shailesh Mishra; Christoph Endres; Thorsten Holz; Mario Fritz On the Hardness of Robustness Transfer: A Perspective from Rademacher Complexity over Symmetric Difference Hypothesis Space. (68%)Yuyang Deng; Nidham Gazagnadou; Junyuan Hong; Mehrdad Mahdavi; Lingjuan Lyu Harnessing the Speed and Accuracy of Machine Learning to Advance Cybersecurity. (2%)Khatoon Mohammed 2023-02-22 Mitigating Adversarial Attacks in Deepfake Detection: An Exploration of Perturbation and AI Techniques. (99%)Saminder Dhesi; Laura Fontes; Pedro Machado; Isibor Kennedy Ihianle; Farhad Fassihi Tash; David Ada Adama PAD: Towards Principled Adversarial Malware Detection Against Evasion Attacks. (98%)Deqiang Li; Shicheng Cui; Yun Li; Jia Xu; Fu Xiao; Shouhuai Xu Provable Robustness Against a Union of $\ell_0$ Adversarial Attacks. (97%)Zayd Hammoudeh; Daniel Lowd ASSET: Robust Backdoor Data Detection Across a Multiplicity of Deep Learning Paradigms. (33%)Minzhou Pan; Yi Zeng; Lingjuan Lyu; Xue Lin; Ruoxi Jia On the Robustness of ChatGPT: An Adversarial and Out-of-distribution Perspective. (12%)Jindong Wang; Xixu Hu; Wenxin Hou; Hao Chen; Runkai Zheng; Yidong Wang; Linyi Yang; Haojun Huang; Wei Ye; Xiubo Geng; Binxin Jiao; Yue Zhang; Xing Xie 2023-02-21 MultiRobustBench: Benchmarking Robustness Against Multiple Attacks. (99%)Sihui Dai; Saeed Mahloujifar; Chong Xiang; Vikash Sehwag; Pin-Yu Chen; Prateek Mittal MalProtect: Stateful Defense Against Adversarial Query Attacks in ML-based Malware Detection. (99%)Aqib Rashid; Jose Such Interpretable Spectrum Transformation Attacks to Speaker Recognition. (98%)Jiadi Yao; Hong Luo; Xiao-Lei Zhang Characterizing the Optimal 0-1 Loss for Multi-class Classification with a Test-time Attacker. (97%)Sihui Dai; Wenxin Ding; Arjun Nitin Bhagoji; Daniel Cullina; Ben Y. Zhao; Haitao Zheng; Prateek Mittal Generalization Bounds for Adversarial Contrastive Learning. (31%)Xin Zou; Weiwei Liu 2023-02-20 An Incremental Gray-box Physical Adversarial Attack on Neural Network Training. (98%)Rabiah Al-qudah; Moayad Aloqaily; Bassem Ouni; Mohsen Guizani; Thierry Lestable Variation Enhanced Attacks Against RRAM-based Neuromorphic Computing System. (97%)Hao Lv; Bing Li; Lei Zhang; Cheng Liu; Ying Wang Seasoning Model Soups for Robustness to Adversarial and Natural Distribution Shifts. (88%)Francesco Croce; Sylvestre-Alvise Rebuffi; Evan Shelhamer; Sven Gowal Poisoning Web-Scale Training Datasets is Practical. (83%)Nicholas Carlini; Matthew Jagielski; Christopher A. Choquette-Choo; Daniel Paleka; Will Pearce; Hyrum Anderson; Andreas Terzis; Kurt Thomas; Florian Tramèr Pseudo Label-Guided Model Inversion Attack via Conditional Generative Adversarial Network. (47%)Xiaojian Yuan; Kejiang Chen; Jie Zhang; Weiming Zhang; Nenghai Yu; Yang Zhang Model-based feature selection for neural networks: A mixed-integer programming approach. (22%)Shudian Zhao; Calvin Tsay; Jan Kronqvist Take Me Home: Reversing Distribution Shifts using Reinforcement Learning. (8%)Vivian Lin; Kuk Jin Jang; Souradeep Dutta; Michele Caprio; Oleg Sokolsky; Insup Lee Prompt Stealing Attacks Against Text-to-Image Generation Models. (1%)Xinyue Shen; Yiting Qu; Michael Backes; Yang Zhang 2023-02-19 X-Adv: Physical Adversarial Object Attacks against X-ray Prohibited Item Detection. (99%)Aishan Liu; Jun Guo; Jiakai Wang; Siyuan Liang; Renshuai Tao; Wenbo Zhou; Cong Liu; Xianglong Liu; Dacheng Tao Stationary Point Losses for Robust Model. (93%)Weiwei Gao; Dazhi Zhang; Yao Li; Zhichang Guo; Ovanes Petrosian On Feasibility of Server-side Backdoor Attacks on Split Learning. (76%)Behrad Tajalli; Oguzhan Ersoy; Stjepan Picek 2023-02-18 Adversarial Machine Learning: A Systematic Survey of Backdoor Attack, Weight Attack and Adversarial Example. (99%)Baoyuan Wu; Li Liu; Zihao Zhu; Qingshan Liu; Zhaofeng He; Siwei Lyu Delving into the Adversarial Robustness of Federated Learning. (98%)Jie Zhang; Bo Li; Chen Chen; Lingjuan Lyu; Shuang Wu; Shouhong Ding; Chao Wu Meta Style Adversarial Training for Cross-Domain Few-Shot Learning. (83%)Yuqian Fu; Yu Xie; Yanwei Fu; Yu-Gang Jiang Towards Safer Generative Language Models: A Survey on Safety Risks, Evaluations, and Improvements. (67%)Jiawen Deng; Jiale Cheng; Hao Sun; Zhexin Zhang; Minlie Huang MedViT: A Robust Vision Transformer for Generalized Medical Image Classification. (12%)Omid Nejati Manzari; Hamid Ahmadabadi; Hossein Kashiani; Shahriar B. Shokouhi; Ahmad Ayatollahi RobustNLP: A Technique to Defend NLP Models Against Backdoor Attacks. (11%)Marwan Omar Beyond Distribution Shift: Spurious Features Through the Lens of Training Dynamics. (2%)Nihal Murali; Aahlad Puli; Ke Yu; Rajesh Ranganath; Kayhan Batmanghelich 2023-02-17 Measuring Equality in Machine Learning Security Defenses. (96%)Luke E. Richards; Edward Raff; Cynthia Matuszek Function Composition in Trustworthy Machine Learning: Implementation Choices, Insights, and Questions. (5%)Manish Nagireddy; Moninder Singh; Samuel C. Hoffman; Evaline Ju; Karthikeyan Natesan Ramamurthy; Kush R. Varshney RetVec: Resilient and Efficient Text Vectorizer. (1%)Elie Bursztein; Marina Zhang; Owen Vallis; Xinyu Jia; Alexey Kurakin 2023-02-16 On the Effect of Adversarial Training Against Invariance-based Adversarial Examples. (99%)Roland Rauter; Martin Nocker; Florian Merkle; Pascal Schöttle High-frequency Matters: An Overwriting Attack and defense for Image-processing Neural Network Watermarking. (67%)Huajie Chen; Tianqing Zhu; Chi Liu; Shui Yu; Wanlei Zhou Marich: A Query-efficient Distributionally Equivalent Model Extraction Attack using Public Data. (3%)Pratik Karmakar; Debabrota Basu A Novel Noise Injection-based Training Scheme for Better Model Robustness. (2%)Zeliang Zhang; Jinyang Jiang; Minjie Chen; Zhiyuan Wang; Yijie Peng; Zhaofei Yu 2023-02-15 Masking and Mixing Adversarial Training. (99%)Hiroki Adachi; Tsubasa Hirakawa; Takayoshi Yamashita; Hironobu Fujiyoshi; Yasunori Ishii; Kazuki Kozuka Robust Mid-Pass Filtering Graph Convolutional Networks. (98%)Jincheng Huang; Lun Du; Xu Chen; Qiang Fu; Shi Han; Dongmei Zhang Graph Adversarial Immunization for Certifiable Robustness. (98%)Shuchang Tao; Huawei Shen; Qi Cao; Yunfan Wu; Liang Hou; Xueqi Cheng XploreNAS: Explore Adversarially Robust & Hardware-efficient Neural Architectures for Non-ideal Xbars. (87%)Abhiroop Bhattacharjee; Abhishek Moitra; Priyadarshini Panda Tight Auditing of Differentially Private Machine Learning. (41%)Milad Nasr; Jamie Hayes; Thomas Steinke; Borja Balle; Florian Tramèr; Matthew Jagielski; Nicholas Carlini; Andreas Terzis Field-sensitive Data Flow Integrity. (1%)So Shizukuishi; Yoshitaka Arahori; Katsuhiko Gondow Uncertainty-Estimation with Normalized Logits for Out-of-Distribution Detection. (1%)Mouxiao Huang; Yu Qiao 2023-02-14 Regret-Based Defense in Adversarial Reinforcement Learning. (99%)Roman Belaire; Pradeep Varakantham; Thanh Nguyen; David Lo On the Role of Randomization in Adversarially Robust Classification. (99%)Lucas Gnecco-Heredia; Yann Chevaleyre; Benjamin Negrevergne; Laurent Meunier; Muni Sreenivas Pydi Attacking Fake News Detectors via Manipulating News Social Engagement. (83%)Haoran Wang; Yingtong Dou; Canyu Chen; Lichao Sun; Philip S. Yu; Kai Shu An Experimental Study of Byzantine-Robust Aggregation Schemes in Federated Learning. (31%)Shenghui Li; Edith C. -H. Ngai; Thiemo Voigt A Modern Look at the Relationship between Sharpness and Generalization. (10%)Maksym Andriushchenko; Francesco Croce; Maximilian Müller; Matthias Hein; Nicolas Flammarion Bounding Training Data Reconstruction in DP-SGD. (8%)Jamie Hayes; Saeed Mahloujifar; Borja Balle Security Defense For Smart Contracts: A Comprehensive Survey. (1%)Nikolay Ivanov; Chenning Li; Qiben Yan; Zhiyuan Sun; Zhichao Cao; Xiapu Luo READIN: A Chinese Multi-Task Benchmark with Realistic and Diverse Input Noises. (1%)Chenglei Si; Zhengyan Zhang; Yingfa Chen; Xiaozhi Wang; Zhiyuan Liu; Maosong Sun 2023-02-13 Sneaky Spikes: Uncovering Stealthy Backdoor Attacks in Spiking Neural Networks with Neuromorphic Data. (98%)Gorka Abad; Oguzhan Ersoy; Stjepan Picek; Aitor Urbieta Raising the Cost of Malicious AI-Powered Image Editing. (82%)Hadi Salman; Alaa Khaddaj; Guillaume Leclerc; Andrew Ilyas; Aleksander Madry Targeted Attack on GPT-Neo for the SATML Language Model Data Extraction Challenge. (8%)Ali Al-Kaswan; Maliheh Izadi; Deursen Arie van Backdoor Learning for NLP: Recent Advances, Challenges, and Future Research Directions. (1%)Marwan Omar 2023-02-12 TextDefense: Adversarial Text Detection based on Word Importance Entropy. (99%)Lujia Shen; Xuhong Zhang; Shouling Ji; Yuwen Pu; Chunpeng Ge; Xing Yang; Yanghe Feng 2023-02-11 Mutation-Based Adversarial Attacks on Neural Text Detectors. (69%)Gongbo Liang; Jesus Guerrero; Izzat Alsmadi HateProof: Are Hateful Meme Detection Systems really Robust? (13%)Piush Aggarwal; Pranit Chawla; Mithun Das; Punyajoy Saha; Binny Mathew; Torsten Zesch; Animesh Mukherjee MTTM: Metamorphic Testing for Textual Content Moderation Software. (2%)Wenxuan Wang; Jen-tse Huang; Weibin Wu; Jianping Zhang; Yizhan Huang; Shuqing Li; Pinjia He; Michael Lyu Pushing the Accuracy-Group Robustness Frontier with Introspective Self-play. (1%)Jeremiah Zhe Liu; Krishnamurthy Dj Dvijotham; Jihyeon Lee; Quan Yuan; Martin Strobel; Balaji Lakshminarayanan; Deepak Ramachandran High Recovery with Fewer Injections: Practical Binary Volumetric Injection Attacks against Dynamic Searchable Encryption. (1%)Xianglong Zhang; Wei Wang; Peng Xu; Laurence T. Yang; Kaitai Liang 2023-02-10 Making Substitute Models More Bayesian Can Enhance Transferability of Adversarial Examples. (98%)Qizhang Li; Yiwen Guo; Wangmeng Zuo; Hao Chen Unnoticeable Backdoor Attacks on Graph Neural Networks. (80%)Enyan Dai; Minhua Lin; Xiang Zhang; Suhang Wang Step by Step Loss Goes Very Far: Multi-Step Quantization for Adversarial Text Attacks. (73%)Piotr Gaiński; Klaudia Bałazy 2023-02-09 IB-RAR: Information Bottleneck as Regularizer for Adversarial Robustness. (98%)Xiaoyun Xu; Guilherme Perin; Stjepan Picek Adversarial Example Does Good: Preventing Painting Imitation from Diffusion Models via Adversarial Examples. (98%)Chumeng Liang; Xiaoyu Wu; Yang Hua; Jiaru Zhang; Yiming Xue; Tao Song; Zhengui Xue; Ruhui Ma; Haibing Guan Mithridates: Auditing and Boosting Backdoor Resistance of Machine Learning Pipelines. (81%)Eugene Bagdasaryan; Vitaly Shmatikov Imperceptible Sample-Specific Backdoor to DNN with Denoising Autoencoder. (62%)Jiliang Zhang; Jing Xu; Zhi Zhang; Yansong Gao Better Diffusion Models Further Improve Adversarial Training. (22%)Zekai Wang; Tianyu Pang; Chao Du; Min Lin; Weiwei Liu; Shuicheng Yan Augmenting NLP data to counter Annotation Artifacts for NLI Tasks. (16%)Armaan Singh Bhullar Incremental Satisfiability Modulo Theory for Verification of Deep Neural Networks. (1%)Pengfei Yang; Zhiming Chi; Zongxin Liu; Mengyu Zhao; Cheng-Chao Huang; Shaowei Cai; Lijun Zhang 2023-02-08 WAT: Improve the Worst-class Robustness in Adversarial Training. (99%)Boqi Li; Weiwei Liu Exploiting Certified Defences to Attack Randomised Smoothing. (99%)Andrew C. Cullen; Paul Montague; Shijie Liu; Sarah M. Erfani; Benjamin I. P. Rubinstein Shortcut Detection with Variational Autoencoders. (13%)Nicolas M. Müller; Simon Roschmann; Shahbaz Khan; Philip Sperl; Konstantin Böttinger Continuous Learning for Android Malware Detection. (13%)Yizheng Chen; Zhoujie Ding; David Wagner Training-free Lexical Backdoor Attacks on Language Models. (8%)Yujin Huang; Terry Yue Zhuo; Qiongkai Xu; Han Hu; Xingliang Yuan; Chunyang Chen On Function-Coupled Watermarks for Deep Neural Networks. (2%)Xiangyu Wen; Yu Li; Wei Jiang; Qiang Xu Unsupervised Learning of Initialization in Deep Neural Networks via Maximum Mean Discrepancy. (1%)Cheolhyoung Lee; Kyunghyun Cho 2023-02-07 Toward Face Biometric De-identification using Adversarial Examples. (98%)Mahdi Ghafourian; Julian Fierrez; Luis Felipe Gomez; Ruben Vera-Rodriguez; Aythami Morales; Zohra Rezgui; Raymond Veldhuis Attacking Cooperative Multi-Agent Reinforcement Learning by Adversarial Minority Influence. (83%)Simin Li; Jun Guo; Jingqiao Xiu; Pu Feng; Xin Yu; Jiakai Wang; Aishan Liu; Wenjun Wu; Xianglong Liu Membership Inference Attacks against Diffusion Models. (64%)Tomoya Matsumoto; Takayuki Miura; Naoto Yanai Temporal Robustness against Data Poisoning. (12%)Wenxiao Wang; Soheil Feizi Robustness Implies Fairness in Casual Algorithmic Recourse. (2%)Ahmad-Reza Ehyaei; Amir-Hossein Karimi; Bernhard Schölkopf; Setareh Maghsudi Low-Latency Communication using Delay-Aware Relays Against Reactive Adversaries. (1%)Vivek Chaudhary; J. Harshan 2023-02-06 Less is More: Understanding Word-level Textual Adversarial Attack via n-gram Frequency Descend. (99%)Ning Lu; Shengcai Liu; Zhirui Zhang; Qi Wang; Haifeng Liu; Ke Tang SCALE-UP: An Efficient Black-box Input-level Backdoor Detection via Analyzing Scaled Prediction Consistency. (92%)Junfeng Guo; Yiming Li; Xun Chen; Hanqing Guo; Lichao Sun; Cong Liu Exploring and Exploiting Decision Boundary Dynamics for Adversarial Robustness. (87%)Yuancheng Xu; Yanchao Sun; Micah Goldblum; Tom Goldstein; Furong Huang Collective Robustness Certificates: Exploiting Interdependence in Graph Neural Networks. (75%)Jan Schuchardt; Aleksandar Bojchevski; Johannes Gasteiger; Stephan Günnemann GAT: Guided Adversarial Training with Pareto-optimal Auxiliary Tasks. (67%)Salah Ghamizi; Jingfeng Zhang; Maxime Cordy; Mike Papadakis; Masashi Sugiyama; Yves Le Traon Target-based Surrogates for Stochastic Optimization. (1%)Jonathan Wilder Lavington; Sharan Vaswani; Reza Babanezhad; Mark Schmidt; Nicolas Le Roux Dropout Injection at Test Time for Post Hoc Uncertainty Quantification in Neural Networks. (1%)Emanuele Ledda; Giorgio Fumera; Fabio Roli One-shot Empirical Privacy Estimation for Federated Learning. (1%)Galen Andrew; Peter Kairouz; Sewoong Oh; Alina Oprea; H. Brendan McMahan; Vinith Suriyakumar 2023-02-05 On the Role of Contrastive Representation Learning in Adversarial Robustness: An Empirical Study. (54%)Fatemeh Ghofrani; Mehdi Yaghouti; Pooyan Jamshidi Leaving Reality to Imagination: Robust Classification via Generated Datasets. (2%)Hritik Bansal; Aditya Grover 2023-02-04 CosPGD: a unified white-box adversarial attack for pixel-wise prediction tasks. (99%)Shashank Agnihotri; Steffen Jung; Margret Keuper A Minimax Approach Against Multi-Armed Adversarial Attacks Detection. (86%)Federica Granese; Marco Romanelli; Siddharth Garg; Pablo Piantanida Run-Off Election: Improved Provable Defense against Data Poisoning Attacks. (83%)Keivan Rezaei; Kiarash Banihashem; Atoosa Chegini; Soheil Feizi AUTOLYCUS: Exploiting Explainable AI (XAI) for Model Extraction Attacks against Decision Tree Models. (80%)Abdullah Caglar Oksuz; Anisa Halimi; Erman Ayday Certified Robust Control under Adversarial Perturbations. (78%)Jinghan Yang; Hunmin Kim; Wenbin Wan; Naira Hovakimyan; Yevgeniy Vorobeychik 2023-02-03 TextShield: Beyond Successfully Detecting Adversarial Sentences in Text Classification. (96%)Lingfeng Shen; Ze Zhang; Haiyun Jiang; Ying Chen DeTorrent: An Adversarial Padding-only Traffic Analysis Defense. (73%)James K Holland; Jason Carpenter; Se Eun Oh; Nicholas Hopper SoK: A Systematic Evaluation of Backdoor Trigger Characteristics in Image Classification. (61%)Gorka Abad; Jing Xu; Stefanos Koffas; Behrad Tajalli; Stjepan Picek; Mauro Conti Beyond the Universal Law of Robustness: Sharper Laws for Random Features and Neural Tangent Kernels. (15%)Simone Bombari; Shayan Kiyani; Marco Mondelli Asymmetric Certified Robustness via Feature-Convex Neural Networks. (8%)Samuel Pfrommer; Brendon G. Anderson; Julien Piet; Somayeh Sojoudi Revisiting Personalized Federated Learning: Robustness Against Backdoor Attacks. (2%)Zeyu Qin; Liuyi Yao; Daoyuan Chen; Yaliang Li; Bolin Ding; Minhao Cheng BarrierBypass: Out-of-Sight Clean Voice Command Injection Attacks through Physical Barriers. (2%)Payton Walker; Tianfang Zhang; Cong Shi; Nitesh Saxena; Yingying Chen From Robustness to Privacy and Back. (2%)Hilal Asi; Jonathan Ullman; Lydia Zakynthinou DCA: Delayed Charging Attack on the Electric Shared Mobility System. (1%)Shuocheng Guo; Hanlin Chen; Mizanur Rahman; Xinwu Qian Augmenting Rule-based DNS Censorship Detection at Scale with Machine Learning. (1%)Jacob Alexander Markson Brown; Xi Jiang; Van Tran; Arjun Nitin Bhagoji; Nguyen Phong Hoang; Nick Feamster; Prateek Mittal; Vinod Yegneswaran 2023-02-02 TransFool: An Adversarial Attack against Neural Machine Translation Models. (99%)Sahar Sadrizadeh; Ljiljana Dolamic; Pascal Frossard Beyond Pretrained Features: Noisy Image Modeling Provides Adversarial Defense. (99%)Zunzhi You; Daochang Liu; Bohyung Han; Chang Xu On the Robustness of Randomized Ensembles to Adversarial Perturbations. (75%)Hassan Dbouk; Naresh R. Shanbhag Provably Bounding Neural Network Preimages. (64%)Suhas Kotha; Christopher Brix; Zico Kolter; Krishnamurthy Dvijotham; Huan Zhang A sliced-Wasserstein distance-based approach for out-of-class-distribution detection. (62%)Mohammad Shifat E Rabbi; Abu Hasnat Mohammad Rubaiyat; Yan Zhuang; Gustavo K Rohde Effective Robustness against Natural Distribution Shifts for Models with Different Training Data. (13%)Zhouxing Shi; Nicholas Carlini; Ananth Balashankar; Ludwig Schmidt; Cho-Jui Hsieh; Alex Beutel; Yao Qin SPECWANDS: An Efficient Priority-based Scheduler Against Speculation Contention Attacks. (10%)Bowen Tang; Chenggang Wu; Pen-Chung Yew; Yinqian Zhang; Mengyao Xie; Yuanming Lai; Yan Kang; Wei Wang; Qiang Wei; Zhe Wang Defensive ML: Defending Architectural Side-channels with Adversarial Obfuscation. (2%)Hyoungwook Nam; Raghavendra Pradyumna Pothukuchi; Bo Li; Nam Sung Kim; Josep Torrellas Generalized Uncertainty of Deep Neural Networks: Taxonomy and Applications. (1%)Chengyu Dong Dataset Distillation Fixes Dataset Reconstruction Attacks. (1%)Noel Loo; Ramin Hasani; Mathias Lechner; Daniela Rus 2023-02-01 Universal Soldier: Using Universal Adversarial Perturbations for Detecting Backdoor Attacks. (99%)Xiaoyun Xu; Oguzhan Ersoy; Stjepan Picek Effectiveness of Moving Target Defenses for Adversarial Attacks in ML-based Malware Detection. (92%)Aqib Rashid; Jose Such Exploring Semantic Perturbations on Grover. (56%)Pranav Kulkarni; Ziqing Ji; Yan Xu; Marko Neskovic; Kevin Nolan BackdoorBox: A Python Toolbox for Backdoor Learning. (10%)Yiming Li; Mengxi Ya; Yang Bai; Yong Jiang; Shu-Tao Xia 2023-01-31 Reverse engineering adversarial attacks with fingerprints from adversarial examples. (99%)David Aaron Embedded Intelligence Nicholson; Vincent Embedded Intelligence Emanuele The Impacts of Unanswerable Questions on the Robustness of Machine Reading Comprehension Models. (97%)Son Quoc Tran; Phong Nguyen-Thuan Do; Uyen Le; Matt Kretchmar Are Defenses for Graph Neural Networks Robust? (80%)Felix Mujkanovic; Simon Geisler; Stephan Günnemann; Aleksandar Bojchevski Adversarial Training of Self-supervised Monocular Depth Estimation against Physical-World Attacks. (75%)Zhiyuan Cheng; James Liang; Guanhong Tao; Dongfang Liu; Xiangyu Zhang Robust Linear Regression: Gradient-descent, Early-stopping, and Beyond. (47%)Meyer Scetbon; Elvis Dohmatob Fairness-aware Vision Transformer via Debiased Self-Attention. (47%)Yao Qiang; Chengyin Li; Prashant Khanduri; Dongxiao Zhu Image Shortcut Squeezing: Countering Perturbative Availability Poisons with Compression. (12%)Zhuoran Liu; Zhengyu Zhao; Martha Larson DRAINCLoG: Detecting Rogue Accounts with Illegally-obtained NFTs using Classifiers Learned on Graphs. (1%)Hanna Kim; Jian Cui; Eugene Jang; Chanhee Lee; Yongjae Lee; Jin-Woo Chung; Seungwon Shin Identifying the Hazard Boundary of ML-enabled Autonomous Systems Using Cooperative Co-Evolutionary Search. (1%)Sepehr Sharifi; Donghwan Shin; Lionel C. Briand; Nathan Aschbacher 2023-01-30 Feature-Space Bayesian Adversarial Learning Improved Malware Detector Robustness. (99%)Bao Gia Doan; Shuiqiao Yang; Paul Montague; Vel Olivier De; Tamas Abraham; Seyit Camtepe; Salil S. Kanhere; Ehsan Abbasnejad; Damith C. Ranasinghe Improving Adversarial Transferability with Scheduled Step Size and Dual Example. (99%)Zeliang Zhang; Peihan Liu; Xiaosen Wang; Chenliang Xu Towards Adversarial Realism and Robust Learning for IoT Intrusion Detection and Classification. (99%)João Vitorino; Isabel Praça; Eva Maia RS-Del: Edit Distance Robustness Certificates for Sequence Classifiers via Randomized Deletion. (99%)Zhuoqun Huang; Neil G. Marchant; Keane Lucas; Lujo Bauer; Olga Ohrimenko; Benjamin I. P. Rubinstein Identifying Adversarially Attackable and Robust Samples. (99%)Vyas Raina; Mark Gales On Robustness of Prompt-based Semantic Parsing with Large Pre-trained Language Model: An Empirical Study on Codex. (98%)Terry Yue Zhuo; Zhuang Li; Yujin Huang; Fatemeh Shiri; Weiqing Wang; Gholamreza Haffari; Yuan-Fang Li Anchor-Based Adversarially Robust Zero-Shot Learning Driven by Language. (96%)Xiao Li; Wei Zhang; Yining Liu; Zhanhao Hu; Bo Zhang; Xiaolin Hu Inference Time Evidences of Adversarial Attacks for Forensic on Transformers. (87%)Hugo Lemarchant; Liangzi Li; Yiming Qian; Yuta Nakashima; Hajime Nagahara On the Efficacy of Metrics to Describe Adversarial Attacks. (82%)Tommaso Puccetti; Tommaso Zoppi; Andrea Ceccarelli Benchmarking Robustness to Adversarial Image Obfuscations. (74%)Florian Stimberg; Ayan Chakrabarti; Chun-Ta Lu; Hussein Hazimeh; Otilia Stretcu; Wei Qiao; Yintao Liu; Merve Kaya; Cyrus Rashtchian; Ariel Fuxman; Mehmet Tek; Sven Gowal Extracting Training Data from Diffusion Models. (5%)Nicholas Carlini; Jamie Hayes; Milad Nasr; Matthew Jagielski; Vikash Sehwag; Florian Tramèr; Borja Balle; Daphne Ippolito; Eric Wallace Affinity Uncertainty-based Hard Negative Mining in Graph Contrastive Learning. (2%)Chaoxi Niu; Guansong Pang; Ling Chen M3FAS: An Accurate and Robust MultiModal Mobile Face Anti-Spoofing System. (1%)Chenqi Kong; Kexin Zheng; Yibing Liu; Shiqi Wang; Anderson Rocha; Haoliang Li 2023-01-29 Unlocking Deterministic Robustness Certification on ImageNet. (98%)Kai Hu; Andy Zou; Zifan Wang; Klas Leino; Matt Fredrikson Mitigating Adversarial Effects of False Data Injection Attacks in Power Grid. (93%)Farhin Farhad Riya; Shahinul Hoque; Jinyuan Stella Sun; Jiangnan Li; Hairong Qi Improving the Accuracy-Robustness Trade-Off of Classifiers via Adaptive Smoothing. (83%)Yatong Bai; Brendon G. Anderson; Aerin Kim; Somayeh Sojoudi Uncovering Adversarial Risks of Test-Time Adaptation. (82%)Tong Wu; Feiran Jia; Xiangyu Qi; Jiachen T. Wang; Vikash Sehwag; Saeed Mahloujifar; Prateek Mittal Adversarial Attacks on Adversarial Bandits. (69%)Yuzhe Ma; Zhijin Zhou Towards Verifying the Geometric Robustness of Large-scale Neural Networks. (54%)Fu Wang; Peipei Xu; Wenjie Ruan; Xiaowei Huang Lateralized Learning for Multi-Class Visual Classification Tasks. (13%)Abubakar Siddique; Will N. Browne; Gina M. Grimshaw Diverse, Difficult, and Odd Instances (D2O): A New Test Set for Object Classification. (3%)Ali Borji Adversarial Style Augmentation for Domain Generalization. (2%)Yabin Zhang; Bin Deng; Ruihuang Li; Kui Jia; Lei Zhang Confidence-Aware Calibration and Scoring Functions for Curriculum Learning. (1%)Shuang Ao; Stefan Rueger; Advaith Siddharthan 2023-01-28 Node Injection for Class-specific Network Poisoning. (82%)Ansh Kumar Sharma; Rahul Kukreja; Mayank Kharbanda; Tanmoy Chakraborty Out-of-distribution Detection with Energy-based Models. (82%)Sven Elflein Gradient Shaping: Enhancing Backdoor Attack Against Reverse Engineering. (13%)Rui Zhu; Di Tang; Siyuan Tang; Guanhong Tao; Shiqing Ma; Xiaofeng Wang; Haixu Tang Selecting Models based on the Risk of Damage Caused by Adversarial Attacks. (1%)Jona Klemenc; Holger Trittenbach 2023-01-27 Semantic Adversarial Attacks on Face Recognition through Significant Attributes. (99%)Yasmeen M. Khedr; Yifeng Xiong; Kun He Targeted Attacks on Timeseries Forecasting. (99%)Yuvaraj Govindarajulu; Avinash Amballa; Pavan Kulkarni; Manojkumar Parmar Adapting Step-size: A Unified Perspective to Analyze and Improve Gradient-based Methods for Adversarial Attacks. (98%)Wei Tao; Lei Bao; Long Sheng; Gaowei Wu; Qing Tao PECAN: A Deterministic Certified Defense Against Backdoor Attacks. (97%)Yuhao Zhang; Aws Albarghouthi; Loris D'Antoni Vertex-based reachability analysis for verifying ReLU deep neural networks. (93%)João Zago; Eduardo Camponogara; Eric Antonelo OccRob: Efficient SMT-Based Occlusion Robustness Verification of Deep Neural Networks. (92%)Xingwu Guo; Ziwei Zhou; Yueling Zhang; Guy Katz; Min Zhang PCV: A Point Cloud-Based Network Verifier. (88%)Arup Kumar Sarker; Farzana Yasmin Ahmad; Matthew B. Dwyer Robust Transformer with Locality Inductive Bias and Feature Normalization. (88%)Omid Nejati Manzari; Hossein Kashiani; Hojat Asgarian Dehkordi; Shahriar Baradaran Shokouhi Analyzing Robustness of the Deep Reinforcement Learning Algorithm in Ramp Metering Applications Considering False Data Injection Attack and Defense. (87%)Diyi Liu; Lanmin Liu; Lee D Han Learning to Unlearn: Instance-wise Unlearning for Pre-trained Classifiers. (80%)Sungmin Cha; Sungjun Cho; Dasol Hwang; Honglak Lee; Taesup Moon; Moontae Lee Certified Invertibility in Neural Networks via Mixed-Integer Programming. (76%)Tianqi Cui; Thomas Bertalan; George J. Pappas; Manfred Morari; Ioannis G. Kevrekidis; Mahyar Fazlyab 2023-01-26 Attacking Important Pixels for Anchor-free Detectors. (99%)Yunxu Xie; Shu Hu; Xin Wang; Quanyu Liao; Bin Zhu; Xi Wu; Siwei Lyu Certified Interpretability Robustness for Class Activation Mapping. (92%)Alex Gu; Tsui-Wei Weng; Pin-Yu Chen; Sijia Liu; Luca Daniel Minerva: A File-Based Ransomware Detector. (68%)Dorjan Hitaj; Giulio Pagnotta; Gaspari Fabio De; Carli Lorenzo De; Luigi V. Mancini Interaction-level Membership Inference Attack Against Federated Recommender Systems. (31%)Wei Yuan; Chaoqun Yang; Quoc Viet Hung Nguyen; Lizhen Cui; Tieke He; Hongzhi Yin 2023-01-25 On the Adversarial Robustness of Camera-based 3D Object Detection. (99%)Shaoyuan Xie; Zichao Li; Zeyu Wang; Cihang Xie RobustPdM: Designing Robust Predictive Maintenance against Adversarial Attacks. (99%)Ayesha Siddique; Ripan Kumar Kundu; Gautam Raj Mode; Khaza Anuarul Hoque BDMMT: Backdoor Sample Detection for Language Models through Model Mutation Testing. (98%)Jiali Wei; Ming Fan; Wenjing Jiao; Wuxia Jin; Ting Liu A Data-Centric Approach for Improving Adversarial Training Through the Lens of Out-of-Distribution Detection. (96%)Mohammad Azizmalayeri; Arman Zarei; Alireza Isavand; Mohammad Taghi Manzuri; Mohammad Hossein Rohban A Study on FGSM Adversarial Training for Neural Retrieval. (75%)Simon Lupart; Stéphane Clinchant Distilling Cognitive Backdoor Patterns within an Image. (5%)Hanxun Huang; Xingjun Ma; Sarah Erfani; James Bailey Connecting metrics for shape-texture knowledge in computer vision. (1%)Tiago Oliveira; Tiago Marques; Arlindo L. Oliveira 2023-01-24 Blockchain-aided Secure Semantic Communication for AI-Generated Content in Metaverse. (13%)Yijing Lin; Hongyang Du; Dusit Niyato; Jiangtian Nie; Jiayi Zhang; Yanyu Cheng; Zhaohui Yang Learning Effective Strategies for Moving Target Defense with Switching Costs. (1%)Vignesh Viswanathan; Megha Bose; Praveen Paruchuri Data Augmentation Alone Can Improve Adversarial Training. (1%)Lin Li; Michael Spratling 2023-01-23 DODEM: DOuble DEfense Mechanism Against Adversarial Attacks Towards Secure Industrial Internet of Things Analytics. (99%)Onat Gungor; Tajana Rosing; Baris Aksanli Practical Adversarial Attacks Against AI-Driven Power Allocation in a Distributed MIMO Network. (92%)Ömer Faruk Tuna; Fehmi Emre Kadan; Leyli Karaçay BayBFed: Bayesian Backdoor Defense for Federated Learning. (78%)Kavita Kumari; Phillip Rieger; Hossein Fereidooni; Murtuza Jadliwala; Ahmad-Reza Sadeghi Backdoor Attacks in Peer-to-Peer Federated Learning. (68%)Gokberk Yar; Cristina Nita-Rotaru; Alina Oprea 2023-01-22 Provable Unrestricted Adversarial Training without Compromise with Generalizability. (99%)Lilin Zhang; Ning Yang; Yanchao Sun; Philip S. Yu ContraBERT: Enhancing Code Pre-trained Models via Contrastive Learning. (8%)Shangqing Liu; Bozhi Wu; Xiaofei Xie; Guozhu Meng; Yang Liu 2023-01-20 Limitations of Piecewise Linearity for Efficient Robustness Certification. (95%)Klas Leino Towards Understanding How Self-training Tolerates Data Backdoor Poisoning. (16%)Soumyadeep Pal; Ren Wang; Yuguang Yao; Sijia Liu Dr.Spider: A Diagnostic Evaluation Benchmark towards Text-to-SQL Robustness. (8%)Shuaichen Chang; Jun Wang; Mingwen Dong; Lin Pan; Henghui Zhu; Alexander Hanbo Li; Wuwei Lan; Sheng Zhang; Jiarong Jiang; Joseph Lilien; Steve Ash; William Yang Wang; Zhiguo Wang; Vittorio Castelli; Patrick Ng; Bing Xiang Defending SDN against packet injection attacks using deep learning. (2%)Anh Tuan Phu; Bo Li; Faheem Ullah; Tanvir Ul Huque; Ranesh Naha; Ali Babar; Hung Nguyen 2023-01-19 On the Vulnerability of Backdoor Defenses for Federated Learning. (62%)Pei Fang; Jinghui Chen On the Relationship Between Information-Theoretic Privacy Metrics And Probabilistic Information Privacy. (31%)Chong Xiao Wang; Wee Peng Tay RNAS-CL: Robust Neural Architecture Search by Cross-Layer Knowledge Distillation. (16%)Utkarsh Nath; Yancheng Wang; Yingzhen Yang Enhancing Deep Learning with Scenario-Based Override Rules: a Case Study. (1%)Adiel Ashrov; Guy Katz 2023-01-17 Denoising Diffusion Probabilistic Models as a Defense against Adversarial Attacks. (98%)Lars Lien Ankile; Anna Midgley; Sebastian Weisshaar Adversarial Robust Deep Reinforcement Learning Requires Redefining Robustness. (68%)Ezgi Korkmaz Label Inference Attack against Split Learning under Regression Setting. (8%)Shangyu Xie; Xin Yang; Yuanshun Yao; Tianyi Liu; Taiqing Wang; Jiankai Sun 2023-01-16 $\beta$-DARTS++: Bi-level Regularization for Proxy-robust Differentiable Architecture Search. (1%)Peng Ye; Tong He; Baopu Li; Tao Chen; Lei Bai; Wanli Ouyang Modeling Uncertain Feature Representation for Domain Generalization. (1%)Xiaotong Li; Zixuan Hu; Jun Liu; Yixiao Ge; Yongxing Dai; Ling-Yu Duan 2023-01-15 BEAGLE: Forensics of Deep Learning Backdoor Attack for Better Defense. (4%)Siyuan Cheng; Guanhong Tao; Yingqi Liu; Shengwei An; Xiangzhe Xu; Shiwei Feng; Guangyu Shen; Kaiyuan Zhang; Qiuling Xu; Shiqing Ma; Xiangyu Zhang 2023-01-14 Adaptive Deep Neural Network Inference Optimization with EENet. (1%)Fatih Ilhan; Ka-Ho Chow; Sihao Hu; Tiansheng Huang; Selim Tekin; Wenqi Wei; Yanzhao Wu; Myungjin Lee; Ramana Kompella; Hugo Latapie; Gaowen Liu; Ling Liu 2023-01-13 On the feasibility of attacking Thai LPR systems with adversarial examples. (99%)Chissanupong Jiamsuchon; Jakapan Suaboot; Norrathep Rattanavipanon 2023-01-12 Security-Aware Approximate Spiking Neural Networks. (87%)Syed Tihaam Ahmad; Ayesha Siddique; Khaza Anuarul Hoque Jamming Attacks on Decentralized Federated Learning in General Multi-Hop Wireless Networks. (3%)Yi Shi; Yalin E. Sagduyu; Tugba Erpek 2023-01-11 Phase-shifted Adversarial Training. (82%)Yeachan Kim; Seongyeon Kim; Ihyeok Seo; Bonggun Shin Universal Detection of Backdoor Attacks via Density-based Clustering and Centroids Analysis. (78%)Wei Guo; Benedetta Tondi; Mauro Barni 2023-01-10 On the Robustness of AlphaFold: A COVID-19 Case Study. (73%)Ismail Alkhouri; Sumit Jha; Andre Beckus; George Atia; Alvaro Velasquez; Rickard Ewetz; Arvind Ramanathan; Susmit Jha CDA: Contrastive-adversarial Domain Adaptation. (38%)Nishant Yadav; Mahbubul Alam; Ahmed Farahat; Dipanjan Ghosh; Chetan Gupta; Auroop R. Ganguly User-Centered Security in Natural Language Processing. (12%)Chris Emmery Leveraging Diffusion For Strong and High Quality Face Morphing Attacks. (3%)Zander W. Blasingame; Chen Liu 2023-01-09 Over-The-Air Adversarial Attacks on Deep Learning Wi-Fi Fingerprinting. (99%)Fei Xiao; Yong Huang; Yingying Zuo; Wei Kuang; Wei Wang On the Susceptibility and Robustness of Time Series Models through Adversarial Attack and Defense. (98%)Asadullah Hill Galib; Bidhan Bashyal Is Federated Learning a Practical PET Yet? (13%)Franziska Boenisch; Adam Dziedzic; Roei Schuster; Ali Shahin Shamsabadi; Ilia Shumailov; Nicolas Papernot SoK: Hardware Defenses Against Speculative Execution Attacks. (1%)Guangyuan Hu; Zecheng He; Ruby Lee 2023-01-08 RobArch: Designing Robust Architectures against Adversarial Attacks. (76%)ShengYun Peng; Weilin Xu; Cory Cornelius; Kevin Li; Rahul Duggal; Duen Horng Chau; Jason Martin MoreauGrad: Sparse and Robust Interpretation of Neural Networks via Moreau Envelope. (1%)Jingwei Zhang; Farzan Farnia 2023-01-07 REaaS: Enabling Adversarially Robust Downstream Classifiers via Robust Encoder as a Service. (99%)Wenjie Qu; Jinyuan Jia; Neil Zhenqiang Gong Adversarial training with informed data selection. (99%)Marcele O. K. Mendonça; Javier Maroto; Pascal Frossard; Paulo S. R. Diniz 2023-01-06 Code Difference Guided Adversarial Example Generation for Deep Code Models. (99%)Zhao Tian; Junjie Chen; Zhi Jin Stealthy Backdoor Attack for Code Models. (98%)Zhou Yang; Bowen Xu; Jie M. Zhang; Hong Jin Kang; Jieke Shi; Junda He; David Lo 2023-01-05 Silent Killer: A Stealthy, Clean-Label, Black-Box Backdoor Attack. (98%)Tzvi Lederer; Gallil Maimon; Lior Rokach gRoMA: a Tool for Measuring the Global Robustness of Deep Neural Networks. (96%)Natan Levy; Raz Yerushalmi; Guy Katz Randomized Message-Interception Smoothing: Gray-box Certificates for Graph Neural Networks. (61%)Yan Scholten; Jan Schuchardt; Simon Geisler; Aleksandar Bojchevski; Stephan Günnemann TrojanPuzzle: Covertly Poisoning Code-Suggestion Models. (4%)Hojjat Aghakhani; Wei Dai; Andre Manoel; Xavier Fernandes; Anant Kharkar; Christopher Kruegel; Giovanni Vigna; David Evans; Ben Zorn; Robert Sim Can Large Language Models Change User Preference Adversarially? (1%)Varshini Subhash 2023-01-04 Availability Adversarial Attack and Countermeasures for Deep Learning-based Load Forecasting. (98%)Wangkun Xu; Fei Teng Beckman Defense. (84%)A. V. Subramanyam GUAP: Graph Universal Attack Through Adversarial Patching. (81%)Xiao Zang; Jie Chen; Bo Yuan Enhancement attacks in biomedical machine learning. (1%)Matthew Rosenblatt; Javid Dadashkarimi; Dustin Scheinost 2023-01-03 Explainability and Robustness of Deep Visual Classification Models. (92%)Jindong Gu Look, Listen, and Attack: Backdoor Attacks Against Video Action Recognition. (83%)Hasan Abed Al Kader Hammoud; Shuming Liu; Mohammed Alkhrashi; Fahad AlBalawi; Bernard Ghanem Backdoor Attacks Against Dataset Distillation. (50%)Yugeng Liu; Zheng Li; Michael Backes; Yun Shen; Yang Zhang Analysis of Label-Flip Poisoning Attack on Machine Learning Based Malware Detector. (33%)Kshitiz Aryal; Maanak Gupta; Mahmoud Abdelsalam 2023-01-02 Efficient Robustness Assessment via Adversarial Spatial-Temporal Focus on Videos. (92%)Wei Xingxing; Wang Songping; Yan Huanqian 2023-01-01 Generalizable Black-Box Adversarial Attack with Meta Learning. (99%)Fei Yin; Yong Zhang; Baoyuan Wu; Yan Feng; Jingyi Zhang; Yanbo Fan; Yujiu Yang ExploreADV: Towards exploratory attack for Neural Networks. (99%)Tianzuo Luo; Yuyi Zhong; Siaucheng Khoo Trojaning semi-supervised learning model via poisoning wild images on the web. (47%)Le Feng; Zhenxing Qian; Sheng Li; Xinpeng Zhang 2022-12-30 Tracing the Origin of Adversarial Attack for Forensic Investigation and Deterrence. (99%)Han Fang; Jiyi Zhang; Yupeng Qiu; Ke Xu; Chengfang Fang; Ee-Chien Chang Guidance Through Surrogate: Towards a Generic Diagnostic Attack. (99%)Muzammal Naseer; Salman Khan; Fatih Porikli; Fahad Shahbaz Khan Defense Against Adversarial Attacks on Audio DeepFake Detection. (91%)Piotr Kawa; Marcin Plata; Piotr Syga Adversarial attacks and defenses on ML- and hardware-based IoT device fingerprinting and identification. (82%)Pedro Miguel Sánchez Sánchez; Alberto Huertas Celdrán; Gérôme Bovet; Gregorio Martínez Pérez Unlearnable Clusters: Towards Label-agnostic Unlearnable Examples. (22%)Jiaming Zhang; Xingjun Ma; Qi Yi; Jitao Sang; Yugang Jiang; Yaowei Wang; Changsheng Xu Targeted k-node Collapse Problem: Towards Understanding the Robustness of Local k-core Structure. (1%)Yuqian Lv; Bo Zhou; Jinhuan Wang; Qi Xuan 2022-12-29 "Real Attackers Don't Compute Gradients": Bridging the Gap Between Adversarial ML Research and Practice. (68%)Giovanni Apruzzese; Hyrum S. Anderson; Savino Dambra; David Freeman; Fabio Pierazzi; Kevin A. Roundy Detection of out-of-distribution samples using binary neuron activation patterns. (11%)Bartlomiej Olber; Krystian Radlak; Adam Popowicz; Michal Szczepankiewicz; Krystian Chachula 2022-12-28 Thermal Heating in ReRAM Crossbar Arrays: Challenges and Solutions. (99%)Kamilya Smagulova; Mohammed E. Fouda; Ahmed Eltawil Certifying Safety in Reinforcement Learning under Adversarial Perturbation Attacks. (98%)Junlin Wu; Hussein Sibai; Yevgeniy Vorobeychik Publishing Efficient On-device Models Increases Adversarial Vulnerability. (95%)Sanghyun Hong; Nicholas Carlini; Alexey Kurakin Differentiable Search of Accurate and Robust Architectures. (92%)Yuwei Ou; Xiangning Xie; Shangce Gao; Yanan Sun; Kay Chen Tan; Jiancheng Lv Robust Ranking Explanations. (76%)Chao Chen; Chenghua Guo; Guixiang Ma; Xi Zhang; Sihong Xie Evaluating Generalizability of Deep Learning Models Using Indian-COVID-19 CT Dataset. (1%)Suba S; Nita Parekh; Ramesh Loganathan; Vikram Pudi; Chinnababu Sunkavalli 2022-12-27 EDoG: Adversarial Edge Detection For Graph Neural Networks. (98%)Xiaojun Xu; Yue Yu; Hanzhang Wang; Alok Lal; Carl A. Gunter; Bo Li Learning When to Use Adaptive Adversarial Image Perturbations against Autonomous Vehicles. (86%)Hyung-Jin Yoon; Hamidreza Jafarnejadsani; Petros Voulgaris Sparse Mixture Once-for-all Adversarial Training for Efficient In-Situ Trade-Off Between Accuracy and Robustness of DNNs. (62%)Souvik Kundu; Sairam Sundaresan; Sharath Nittur Sridhar; Shunlin Lu; Han Tang; Peter A. Beerel XMAM:X-raying Models with A Matrix to Reveal Backdoor Attacks for Federated Learning. (56%)Jianyi Zhang; Fangjiao Zhang; Qichao Jin; Zhiqiang Wang; Xiaodong Lin; Xiali Hei 2022-12-25 Simultaneously Optimizing Perturbations and Positions for Black-box Adversarial Patch Attacks. (99%)Xingxing Wei; Ying Guo; Jie Yu; Bo Zhang 2022-12-24 Frequency Regularization for Improving Adversarial Robustness. (99%)Binxiao Huang; Chaofan Tao; Rui Lin; Ngai Wong 2022-12-23 Out-of-Distribution Detection with Reconstruction Error and Typicality-based Penalty. (61%)Genki Osada; Takahashi Tsubasa; Budrul Ahsan; Takashi Nishide Towards Scalable Physically Consistent Neural Networks: an Application to Data-driven Multi-zone Thermal Building Models. (1%)Natale Loris Di; Bratislav Svetozarevic; Philipp Heer; Colin Neil Jones 2022-12-22 Adversarial Machine Learning and Defense Game for NextG Signal Classification with Deep Learning. (98%)Yalin E. Sagduyu Aliasing is a Driver of Adversarial Attacks. (80%)Adrián Rodríguez-Muñoz; Antonio Torralba GAN-based Domain Inference Attack. (2%)Yuechun Gu; Keke Chen Hybrid Quantum-Classical Generative Adversarial Network for High Resolution Image Generation. (1%)Shu Lok Tsang; Maxwell T. West; Sarah M. Erfani; Muhammad Usman 2022-12-21 Revisiting Residual Networks for Adversarial Robustness: An Architectural Perspective. (80%)Shihua Huang; Zhichao Lu; Kalyanmoy Deb; Vishnu Naresh Boddeti Vulnerabilities of Deep Learning-Driven Semantic Communications to Backdoor (Trojan) Attacks. (67%)Yalin E. Sagduyu; Tugba Erpek; Sennur Ulukus; Aylin Yener A Theoretical Study of The Effects of Adversarial Attacks on Sparse Regression. (13%)Deepak Maurya; Jean Honorio 2022-12-20 A Comprehensive Study and Comparison of the Robustness of 3D Object Detectors Against Adversarial Attacks. (98%)Yifan Zhang; Junhui Hou; Yixuan Yuan Multi-head Uncertainty Inference for Adversarial Attack Detection. (98%)Yuqi Yang; Songyun Yang; Jiyang Xie. Zhongwei Si; Kai Guo; Ke Zhang; Kongming Liang In and Out-of-Domain Text Adversarial Robustness via Label Smoothing. (98%)Yahan Yang; Soham Dan; Dan Roth; Insup Lee Is Semantic Communications Secure? A Tale of Multi-Domain Adversarial Attacks. (96%)Yalin E. Sagduyu; Tugba Erpek; Sennur Ulukus; Aylin Yener Unleashing the Power of Visual Prompting At the Pixel Level. (92%)Junyang Wu; Xianhang Li; Chen Wei; Huiyu Wang; Alan Yuille; Yuyin Zhou; Cihang Xie Learned Systems Security. (78%)Roei Schuster; Jin Peng Zhou; Paul Grubbs; Thorsten Eisenhofer; Nicolas Papernot Hidden Poison: Machine Unlearning Enables Camouflaged Poisoning Attacks. (22%)Jimmy Z. Di; Jack Douglas; Jayadev Acharya; Gautam Kamath; Ayush Sekhari ReCode: Robustness Evaluation of Code Generation Models. (10%)Shiqi Wang; Zheng Li; Haifeng Qian; Chenghao Yang; Zijian Wang; Mingyue Shang; Varun Kumar; Samson Tan; Baishakhi Ray; Parminder Bhatia; Ramesh Nallapati; Murali Krishna Ramanathan; Dan Roth; Bing Xiang Defending Against Poisoning Attacks in Open-Domain Question Answering. (8%)Orion Weller; Aleem Khan; Nathaniel Weir; Dawn Lawrie; Durme Benjamin Van SoK: Analysis of Root Causes and Defense Strategies for Attacks on Microarchitectural Optimizations. (5%)Nadja Ramhöj Holtryd; Madhavan Manivannan; Per Stenström DISCO: Distilling Phrasal Counterfactuals with Large Language Models. (1%)Zeming Chen; Qiyue Gao; Kyle Richardson; Antoine Bosselut; Ashish Sabharwal 2022-12-19 TextGrad: Advancing Robustness Evaluation in NLP by Gradient-Driven Optimization. (99%)Bairu Hou; Jinghan Jia; Yihua Zhang; Guanhua Zhang; Yang Zhang; Sijia Liu; Shiyu Chang Towards Robustness of Text-to-SQL Models Against Natural and Realistic Adversarial Table Perturbation. (75%)Xinyu Pi; Bing Wang; Yan Gao; Jiaqi Guo; Zhoujun Li; Jian-Guang Lou AI Security for Geoscience and Remote Sensing: Challenges and Future Trends. (50%)Yonghao Xu; Tao Bai; Weikang Yu; Shizhen Chang; Peter M. Atkinson; Pedram Ghamisi Task-Oriented Communications for NextG: End-to-End Deep Learning and AI Security Aspects. (26%)Yalin E. Sagduyu; Sennur Ulukus; Aylin Yener Flareon: Stealthy any2any Backdoor Injection via Poisoned Augmentation. (2%)Tianrui Qin; Xianghuan He; Xitong Gao; Yiren Zhao; Kejiang Ye; Cheng-Zhong Xu Exploring Optimal Substructure for Out-of-distribution Generalization via Feature-targeted Model Pruning. (1%)Yingchun Wang; Jingcai Guo; Song Guo; Weizhan Zhang; Jie Zhang 2022-12-18 Estimating the Adversarial Robustness of Attributions in Text with Transformers. (99%)Adam Ivankay; Mattia Rigotti; Ivan Girardi; Chiara Marchiori; Pascal Frossard Minimizing Maximum Model Discrepancy for Transferable Black-box Targeted Attacks. (99%)Anqi Zhao; Tong Chu; Yahao Liu; Wen Li; Jingjing Li; Lixin Duan Discrete Point-wise Attack Is Not Enough: Generalized Manifold Adversarial Attack for Face Recognition. (99%)Qian Li; Yuxiao Hu; Ye Liu; Dongxiao Zhang; Xin Jin; Yuntian Chen Fine-Tuning Is All You Need to Mitigate Backdoor Attacks. (4%)Zeyang Sha; Xinlei He; Pascal Berrang; Mathias Humbert; Yang Zhang 2022-12-17 Confidence-aware Training of Smoothed Classifiers for Certified Robustness. (86%)Jongheon Jeong; Seojin Kim; Jinwoo Shin A Review of Speech-centric Trustworthy Machine Learning: Privacy, Safety, and Fairness. (2%)Tiantian Feng; Rajat Hebbar; Nicholas Mehlman; Xuan Shi; Aditya Kommineni; and Shrikanth Narayanan HyPe: Better Pre-trained Language Model Fine-tuning with Hidden Representation Perturbation. (1%)Hongyi Yuan; Zheng Yuan; Chuanqi Tan; Fei Huang; Songfang Huang 2022-12-16 Adversarial Example Defense via Perturbation Grading Strategy. (99%)Shaowei Zhu; Wanli Lyu; Bin Li; Zhaoxia Yin; Bin Luo WebAssembly Diversification for Malware Evasion. (5%)Javier Cabrera-Arteaga; Martin Monperrus; Tim Toady; Benoit Baudry Biomedical image analysis competitions: The state of current participation practice. (4%)Matthias Eisenmann; Annika Reinke; Vivienn Weru; Minu Dietlinde Tizabi; Fabian Isensee; Tim J. Adler; Patrick Godau; Veronika Cheplygina; Michal Kozubek; Sharib Ali; Anubha Gupta; Jan Kybic; Alison Noble; Solórzano Carlos Ortiz de; Samiksha Pachade; Caroline Petitjean; Daniel Sage; Donglai Wei; Elizabeth Wilden; Deepak Alapatt; Vincent Andrearczyk; Ujjwal Baid; Spyridon Bakas; Niranjan Balu; Sophia Bano; Vivek Singh Bawa; Jorge Bernal; Sebastian Bodenstedt; Alessandro Casella; Jinwook Choi; Olivier Commowick; Marie Daum; Adrien Depeursinge; Reuben Dorent; Jan Egger; Hannah Eichhorn; Sandy Engelhardt; Melanie Ganz; Gabriel Girard; Lasse Hansen; Mattias Heinrich; Nicholas Heller; Alessa Hering; Arnaud Huaulmé; Hyunjeong Kim; Bennett Landman; Hongwei Bran Li; Jianning Li; Jun Ma; Anne Martel; Carlos Martín-Isla; Bjoern Menze; Chinedu Innocent Nwoye; Valentin Oreiller; Nicolas Padoy; Sarthak Pati; Kelly Payette; Carole Sudre; Wijnen Kimberlin van; Armine Vardazaryan; Tom Vercauteren; Martin Wagner; Chuanbo Wang; Moi Hoon Yap; Zeyun Yu; Chun Yuan; Maximilian Zenk; Aneeq Zia; David Zimmerer; Rina Bao; Chanyeol Choi; Andrew Cohen; Oleh Dzyubachyk; Adrian Galdran; Tianyuan Gan; Tianqi Guo; Pradyumna Gupta; Mahmood Haithami; Edward Ho; Ikbeom Jang; Zhili Li; Zhengbo Luo; Filip Lux; Sokratis Makrogiannis; Dominik Müller; Young-tack Oh; Subeen Pang; Constantin Pape; Gorkem Polat; Charlotte Rosalie Reed; Kanghyun Ryu; Tim Scherr; Vajira Thambawita; Haoyu Wang; Xinliang Wang; Kele Xu; Hung Yeh; Doyeob Yeo; Yixuan Yuan; Yan Zeng; Xin Zhao; Julian Abbing; Jannes Adam; Nagesh Adluru; Niklas Agethen; Salman Ahmed; Yasmina Al Khalil; Mireia Alenyà; Esa Alhoniemi; Chengyang An; Talha Anwar; Tewodros Weldebirhan Arega; Netanell Avisdris; Dogu Baran Aydogan; Yingbin Bai; Maria Baldeon Calisto; Berke Doga Basaran; Marcel Beetz; Cheng Bian; Hao Bian; Kevin Blansit; Louise Bloch; Robert Bohnsack; Sara Bosticardo; Jack Breen; Mikael Brudfors; Raphael Brüngel; Mariano Cabezas; Alberto Cacciola; Zhiwei Chen; Yucong Chen; Daniel Tianming Chen; Minjeong Cho; Min-Kook Choi; Chuantao Xie Chuantao Xie; Dana Cobzas; Julien Cohen-Adad; Jorge Corral Acero; Sujit Kumar Das; Oliveira Marcela de; Hanqiu Deng; Guiming Dong; Lars Doorenbos; Cory Efird; Di Fan; Mehdi Fatan Serj; Alexandre Fenneteau; Lucas Fidon; Patryk Filipiak; René Finzel; Nuno R. Freitas; Christoph M. Friedrich; Mitchell Fulton; Finn Gaida; Francesco Galati; Christoforos Galazis; Chang Hee Gan; Zheyao Gao; Shengbo Gao; Matej Gazda; Beerend Gerats; Neil Getty; Adam Gibicar; Ryan Gifford; Sajan Gohil; Maria Grammatikopoulou; Daniel Grzech; Orhun Güley; Timo Günnemann; Chunxu Guo; Sylvain Guy; Heonjin Ha; Luyi Han; Il Song Han; Ali Hatamizadeh; Tian He; Jimin Heo; Sebastian Hitziger; SeulGi Hong; SeungBum Hong; Rian Huang; Ziyan Huang; Markus Huellebrand; Stephan Huschauer; Mustaffa Hussain; Tomoo Inubushi; Ece Isik Polat; Mojtaba Jafaritadi; SeongHun Jeong; Bailiang Jian; Yuanhong Jiang; Zhifan Jiang; Yueming Jin; Smriti Joshi; Abdolrahim Kadkhodamohammadi; Reda Abdellah Kamraoui; Inha Kang; Junghwa Kang; Davood Karimi; April Khademi; Muhammad Irfan Khan; Suleiman A. Khan; Rishab Khantwal; Kwang-Ju Kim; Timothy Kline; Satoshi Kondo; Elina Kontio; Adrian Krenzer; Artem Kroviakov; Hugo Kuijf; Satyadwyoom Kumar; Rosa Francesco La; Abhi Lad; Doohee Lee; Minho Lee; Chiara Lena; Hao Li; Ling Li; Xingyu Li; Fuyuan Liao; KuanLun Liao; Arlindo Limede Oliveira; Chaonan Lin; Shan Lin; Akis Linardos; Marius George Linguraru; Han Liu; Tao Liu; Di Liu; Yanling Liu; João Lourenço-Silva; Jingpei Lu; Jiangshan Lu; Imanol Luengo; Christina B. Lund; Huan Minh Luu; Yi Lv; Yi Lv; Uzay Macar; Leon Maechler; Sina Mansour L.; Kenji Marshall; Moona Mazher; Richard McKinley; Alfonso Medela; Felix Meissen; Mingyuan Meng; Dylan Miller; Seyed Hossein Mirjahanmardi; Arnab Mishra; Samir Mitha; Hassan Mohy-ud-Din; Tony Chi Wing Mok; Gowtham Krishnan Murugesan; Enamundram Naga Karthik; Sahil Nalawade; Jakub Nalepa; Mohamed Naser; Ramin Nateghi; Hammad Naveed; Quang-Minh Nguyen; Cuong Nguyen Quoc; Brennan Nichyporuk; Bruno Oliveira; David Owen; Jimut Bahan Pal; Junwen Pan; Wentao Pan; Winnie Pang; Bogyu Park; Vivek Pawar; Kamlesh Pawar; Michael Peven; Lena Philipp; Tomasz Pieciak; Szymon Plotka; Marcel Plutat; Fattaneh Pourakpour; Domen Preložnik; Kumaradevan Punithakumar; Abdul Qayyum; Sandro Queirós; Arman Rahmim; Salar Razavi; Jintao Ren; Mina Rezaei; Jonathan Adam Rico; ZunHyan Rieu; Markus Rink; Johannes Roth; Yusely Ruiz-Gonzalez; Numan Saeed; Anindo Saha; Mostafa Salem; Ricardo Sanchez-Matilla; Kurt Schilling; Wei Shao; Zhiqiang Shen; Ruize Shi; Pengcheng Shi; Daniel Sobotka; Théodore Soulier; Bella Specktor Fadida; Danail Stoyanov; Timothy Sum Hon Mun; Xiaowu Sun; Rong Tao; Franz Thaler; Antoine Théberge; Felix Thielke; Helena Torres; Kareem A. Wahid; Jiacheng Wang; YiFei Wang; Wei Wang; Xiong Wang; Jianhui Wen; Ning Wen; Marek Wodzinski; Ye Wu; Fangfang Xia; Tianqi Xiang; Chen Xiaofei; Lizhan Xu; Tingting Xue; Yuxuan Yang; Lin Yang; Kai Yao; Huifeng Yao; Amirsaeed Yazdani; Michael Yip; Hwanseung Yoo; Fereshteh Yousefirizi; Shunkai Yu; Lei Yu; Jonathan Zamora; Ramy Ashraf Zeineldin; Dewen Zeng; Jianpeng Zhang; Bokai Zhang; Jiapeng Zhang; Fan Zhang; Huahong Zhang; Zhongchen Zhao; Zixuan Zhao; Jiachen Zhao; Can Zhao; Qingshuo Zheng; Yuheng Zhi; Ziqi Zhou; Baosheng Zou; Klaus Maier-Hein; Paul F. Jäger; Annette Kopp-Schneider; Lena Maier-Hein Better May Not Be Fairer: Can Data Augmentation Mitigate Subgroup Degradation? (1%)Ming-Chang Chiu; Pin-Yu Chen; Xuezhe Ma On Human Visual Contrast Sensitivity and Machine Vision Robustness: A Comparative Study. (1%)Ming-Chang Chiu; Yingfei Wang; Derrick Eui Gyu Kim; Pin-Yu Chen; Xuezhe Ma 2022-12-15 Alternating Objectives Generates Stronger PGD-Based Adversarial Attacks. (98%)Nikolaos Antoniou; Efthymios Georgiou; Alexandros Potamianos On Evaluating Adversarial Robustness of Chest X-ray Classification: Pitfalls and Best Practices. (84%)Salah Ghamizi; Maxime Cordy; Michail Papadakis; Yves Le Traon Are Multimodal Models Robust to Image and Text Perturbations? (5%)Jielin Qiu; Yi Zhu; Xingjian Shi; Florian Wenzel; Zhiqiang Tang; Ding Zhao; Bo Li; Mu Li Holistic risk assessment of inference attacks in machine learning. (4%)Yang Yang Defending against cybersecurity threats to the payments and banking system. (2%)Williams Haruna; Toyin Ajiboro Aremu; Yetunde Ajao Modupe White-box Inference Attacks against Centralized Machine Learning and Federated Learning. (1%)Jingyi Ge 2022-12-14 SAIF: Sparse Adversarial and Interpretable Attack Framework. (99%)Tooba Imtiaz; Morgan Kohler; Jared Miller; Zifeng Wang; Mario Sznaier; Octavia Camps; Jennifer Dy Dissecting Distribution Inference. (88%)Anshuman Suri; Yifu Lu; Yanjin Chen; David Evans Generative Robust Classification. (11%)Xuwang Yin Synthesis of Adversarial DDOS Attacks Using Tabular Generative Adversarial Networks. (8%)Abdelmageed Ahmed Hassan; Mohamed Sayed Hussein; Ahmed Shehata AboMoustafa; Sarah Hossam Elmowafy DOC-NAD: A Hybrid Deep One-class Classifier for Network Anomaly Detection. (1%)Mohanad Sarhan; Gayan Kulatilleke; Wai Weng Lo; Siamak Layeghy; Marius Portmann 2022-12-13 Object-fabrication Targeted Attack for Object Detection. (99%)Xuchong Zhang; Changfeng Sun; Haoliang Han; Hang Wang; Hongbin Sun; Nanning Zheng Adversarial Attacks and Defences for Skin Cancer Classification. (99%)Vinay Jogani; Joy Purohit; Ishaan Shivhare; Samina Attari; Shraddha Surtkar Unfolding Local Growth Rate Estimates for (Almost) Perfect Adversarial Detection. (99%)Peter Lorenz; Margret Keuper; Janis Keuper Towards Efficient and Domain-Agnostic Evasion Attack with High-dimensional Categorical Inputs. (80%)Hongyan Bao; Yufei Han; Yujun Zhou; Xin Gao; Xiangliang Zhang Understanding Zero-Shot Adversarial Robustness for Large-Scale Models. (73%)Chengzhi Mao; Scott Geng; Junfeng Yang; Xin Wang; Carl Vondrick Pixel is All You Need: Adversarial Trajectory-Ensemble Active Learning for Salient Object Detection. (56%)Zhenyu Wu; Lin Wang; Wei Wang; Qing Xia; Chenglizhao Chen; Aimin Hao; Shuo Li AdvCat: Domain-Agnostic Robustness Assessment for Cybersecurity-Critical Applications with Categorical Inputs. (56%)Helene Orsini; Hongyan Bao; Yujun Zhou; Xiangrui Xu; Yufei Han; Longyang Yi; Wei Wang; Xin Gao; Xiangliang Zhang Privacy-preserving Security Inference Towards Cloud-Edge Collaborative Using Differential Privacy. (1%)Yulong Wang; Xingshu Chen; Qixu Wang Boosting Semi-Supervised Learning with Contrastive Complementary Labeling. (1%)Qinyi Deng; Yong Guo; Zhibang Yang; Haolin Pan; Jian Chen 2022-12-12 SRoUDA: Meta Self-training for Robust Unsupervised Domain Adaptation. (98%)Wanqing Zhu; Jia-Li Yin; Bo-Hao Chen; Ximeng Liu Adversarially Robust Video Perception by Seeing Motion. (98%)Lingyu Zhang; Chengzhi Mao; Junfeng Yang; Carl Vondrick A Survey on Reinforcement Learning Security with Application to Autonomous Driving. (96%)Ambra Demontis; Maura Pintor; Luca Demetrio; Kathrin Grosse; Hsiao-Ying Lin; Chengfang Fang; Battista Biggio; Fabio Roli HOTCOLD Block: Fooling Thermal Infrared Detectors with a Novel Wearable Design. (96%)Hui Wei; Zhixiang Wang; Xuemei Jia; Yinqiang Zheng; Hao Tang; Shin'ichi Satoh; Zheng Wang Robust Perception through Equivariance. (96%)Chengzhi Mao; Lingyu Zhang; Abhishek Joshi; Junfeng Yang; Hao Wang; Carl Vondrick Despite "super-human" performance, current LLMs are unsuited for decisions about ethics and safety. (75%)Joshua Albrecht; Ellie Kitanidis; Abraham J. Fetterman AFLGuard: Byzantine-robust Asynchronous Federated Learning. (15%)Minghong Fang; Jia Liu; Neil Zhenqiang Gong; Elizabeth S. Bentley Carpet-bombing patch: attacking a deep network without usual requirements. (2%)Pol Labarbarie; Adrien Chan-Hon-Tong; Stéphane Herbin; Milad Leyli-Abadi Numerical Stability of DeepGOPlus Inference. (1%)Inés Gonzalez Pepe; Yohan Chatelain; Gregory Kiar; Tristan Glatard 2022-12-11 DISCO: Adversarial Defense with Local Implicit Functions. (99%)Chih-Hui Ho; Nuno Vasconcelos REAP: A Large-Scale Realistic Adversarial Patch Benchmark. (98%)Nabeel Hingun; Chawin Sitawarin; Jerry Li; David Wagner 2022-12-10 General Adversarial Defense Against Black-box Attacks via Pixel Level and Feature Level Distribution Alignments. (99%)Xiaogang Xu; Hengshuang Zhao; Philip Torr; Jiaya Jia Untargeted Attack against Federated Recommendation Systems via Poisonous Item Embeddings and the Defense. (93%)Yang Yu; Qi Liu; Likang Wu; Runlong Yu; Sanshi Lei Yu; Zaixi Zhang Targeted Adversarial Attacks on Deep Reinforcement Learning Policies via Model Checking. (93%)Dennis Gross; Thiago D. Simao; Nils Jansen; Guillermo A. Perez Mitigating Adversarial Gray-Box Attacks Against Phishing Detectors. (54%)Giovanni Apruzzese; V. S. Subrahmanian How to Backdoor Diffusion Models? (12%)Sheng-Yen Chou; Pin-Yu Chen; Tsung-Yi Ho Identifying the Source of Vulnerability in Explanation Discrepancy: A Case Study in Neural Text Classification. (1%)Ruixuan Tang; Hanjie Chen; Yangfeng Ji 2022-12-09 Understanding and Combating Robust Overfitting via Input Loss Landscape Analysis and Regularization. (98%)Lin Li; Michael Spratling Expeditious Saliency-guided Mix-up through Random Gradient Thresholding. (2%)Minh-Long Luu; Zeyi Huang; Eric P. Xing; Yong Jae Lee; Haohan Wang Spurious Features Everywhere -- Large-Scale Detection of Harmful Spurious Features in ImageNet. (1%)Yannic Neuhaus; Maximilian Augustin; Valentyn Boreiko; Matthias Hein Robustness Implies Privacy in Statistical Estimation. (1%)Samuel B. Hopkins; Gautam Kamath; Mahbod Majid; Shyam Narayanan Selective Amnesia: On Efficient, High-Fidelity and Blind Suppression of Backdoor Effects in Trojaned Machine Learning Models. (1%)Rui Zhu; Di Tang; Siyuan Tang; XiaoFeng Wang; Haixu Tang QVIP: An ILP-based Formal Verification Approach for Quantized Neural Networks. (1%)Yedi Zhang; Zhe Zhao; Fu Song; Min Zhang; Taolue Chen; Jun Sun 2022-12-08 Targeted Adversarial Attacks against Neural Network Trajectory Predictors. (99%)Kaiyuan Tan; Jun Wang; Yiannis Kantaros XRand: Differentially Private Defense against Explanation-Guided Attacks. (68%)Truc Nguyen; Phung Lai; NhatHai Phan; My T. Thai Robust Graph Representation Learning via Predictive Coding. (22%)Billy Byiringiro; Tommaso Salvatori; Thomas Lukasiewicz 2022-12-07 Use of Cryptography in Malware Obfuscation. (1%)Hassan Jameel Asghar; Benjamin Zi Hao Zhao; Muhammad Ikram; Giang Nguyen; Dali Kaafar; Sean Lamont; Daniel Coscia 2022-12-06 Pre-trained Encoders in Self-Supervised Learning Improve Secure and Privacy-preserving Supervised Learning. (96%)Hongbin Liu; Wenjie Qu; Jinyuan Jia; Neil Zhenqiang Gong 2022-12-05 Enhancing Quantum Adversarial Robustness by Randomized Encodings. (99%)Weiyuan Gong; Dong Yuan; Weikang Li; Dong-Ling Deng Multiple Perturbation Attack: Attack Pixelwise Under Different $\ell_p$-norms For Better Adversarial Performance. (99%)Ngoc N. Tran; Anh Tuan Bui; Dinh Phung; Trung Le FaceQAN: Face Image Quality Assessment Through Adversarial Noise Exploration. (92%)Žiga Babnik; Peter Peer; Vitomir Štruc Refiner: Data Refining against Gradient Leakage Attacks in Federated Learning. (76%)Mingyuan Fan; Cen Chen; Chengyu Wang; Wenmeng Zhou; Jun Huang; Ximeng Liu; Wenzhong Guo Blessings and Curses of Covariate Shifts: Adversarial Learning Dynamics, Directional Convergence, and Equilibria. (8%)Tengyuan Liang What is the Solution for State-Adversarial Multi-Agent Reinforcement Learning? (3%)Songyang Han; Sanbao Su; Sihong He; Shuo Han; Haizhao Yang; Fei Miao Spuriosity Rankings: Sorting Data for Spurious Correlation Robustness. (1%)Mazda Moayeri; Wenxiao Wang; Sahil Singla; Soheil Feizi Efficient Malware Analysis Using Metric Embeddings. (1%)Ethan M. Rudd; David Krisiloff; Scott Coull; Daniel Olszewski; Edward Raff; James Holt 2022-12-04 Bayesian Learning with Information Gain Provably Bounds Risk for a Robust Adversarial Defense. (98%)Bao Gia Doan; Ehsan Abbasnejad; Javen Qinfeng Shi; Damith C. Ranasinghe Recognizing Object by Components with Human Prior Knowledge Enhances Adversarial Robustness of Deep Neural Networks. (88%)Xiao Li; Ziqi Wang; Bo Zhang; Fuchun Sun; Xiaolin Hu CSTAR: Towards Compact and STructured Deep Neural Networks with Adversarial Robustness. (82%)Huy Phan; Miao Yin; Yang Sui; Bo Yuan; Saman Zonouz FedCC: Robust Federated Learning against Model Poisoning Attacks. (45%)Hyejun Jeong; Hamin Son; Seohu Lee; Jayun Hyun; Tai-Myoung Chung ConfounderGAN: Protecting Image Data Privacy with Causal Confounder. (8%)Qi Tian; Kun Kuang; Kelu Jiang; Furui Liu; Zhihua Wang; Fei Wu 2022-12-03 LDL: A Defense for Label-Based Membership Inference Attacks. (83%)Arezoo Rajabi; Dinuka Sahabandu; Luyao Niu; Bhaskar Ramasubramanian; Radha Poovendran Security Analysis of SplitFed Learning. (8%)Momin Ahmad Khan; Virat Shejwalkar; Amir Houmansadr; Fatima Muhammad Anwar 2022-12-02 Membership Inference Attacks Against Semantic Segmentation Models. (45%)Tomas Chobola; Dmitrii Usynin; Georgios Kaissis Guaranteed Conformance of Neurosymbolic Models to Natural Constraints. (1%)Kaustubh Sridhar; Souradeep Dutta; James Weimer; Insup Lee 2022-12-01 Purifier: Defending Data Inference Attacks via Transforming Confidence Scores. (89%)Ziqi Yang; Lijin Wang; Da Yang; Jie Wan; Ziming Zhao; Ee-Chien Chang; Fan Zhang; Kui Ren Pareto Regret Analyses in Multi-objective Multi-armed Bandit. (41%)Mengfan Xu; Diego Klabjan All You Need Is Hashing: Defending Against Data Reconstruction Attack in Vertical Federated Learning. (3%)Pengyu Qiu; Xuhong Zhang; Shouling Ji; Yuwen Pu; Ting Wang Generalizing and Improving Jacobian and Hessian Regularization. (1%)Chenwei Cui; Zehao Yan; Guangshen Liu; Liangfu Lu On the Limit of Explaining Black-box Temporal Graph Neural Networks. (1%)Minh N. Vu; My T. Thai SimpleMind adds thinking to deep neural networks. (1%)Youngwon Choi; M. Wasil Wahi-Anwar; Matthew S. Brown 2022-11-30 Towards Interpreting Vulnerability of Multi-Instance Learning via Customized and Universal Adversarial Perturbations. (97%)Yu-Xuan Zhang; Hua Meng; Xue-Mei Cao; Zhengchun Zhou; Mei Yang; Avik Ranjan Adhikary Interpretation of Neural Networks is Susceptible to Universal Adversarial Perturbations. (84%)Haniyeh Ehsani Oskouie; Farzan Farnia Efficient Adversarial Input Generation via Neural Net Patching. (75%)Tooba Khan; Kumar Madhukar; Subodh Vishnu Sharma Toward Robust Diagnosis: A Contour Attention Preserving Adversarial Defense for COVID-19 Detection. (69%)Kun Xiang; Xing Zhang; Jinwen She; Jinpeng Liu; Haohan Wang; Shiqi Deng; Shancheng Jiang Tight Certification of Adversarially Trained Neural Networks via Nonconvex Low-Rank Semidefinite Relaxations. (38%)Hong-Ming Chiu; Richard Y. Zhang Improved Smoothed Analysis of 2-Opt for the Euclidean TSP. (8%)Bodo Manthey; Rhijn Jesse van 2022-11-29 Understanding and Enhancing Robustness of Concept-based Models. (99%)Sanchit Sinha; Mengdi Huai; Jianhui Sun; Aidong Zhang Ada3Diff: Defending against 3D Adversarial Point Clouds via Adaptive Diffusion. (99%)Kui Zhang; Hang Zhou; Jie Zhang; Qidong Huang; Weiming Zhang; Nenghai Yu Advancing Deep Metric Learning Through Multiple Batch Norms And Multi-Targeted Adversarial Examples. (88%)Inderjeet Singh; Kazuya Kakizaki; Toshinori Araki Penalizing Confident Predictions on Largely Perturbed Inputs Does Not Improve Out-of-Distribution Generalization in Question Answering. (83%)Kazutoshi Shinoda; Saku Sugawara; Akiko Aizawa Quantization-aware Interval Bound Propagation for Training Certifiably Robust Quantized Neural Networks. (73%)Mathias Lechner; Đorđe Žikelić; Krishnendu Chatterjee; Thomas A. Henzinger; Daniela Rus AdvMask: A Sparse Adversarial Attack Based Data Augmentation Method for Image Classification. (54%)Suorong Yang; Jinqiao Li; Jian Zhao; Furao Shen A3T: Accuracy Aware Adversarial Training. (10%)Enes Altinisik; Safa Messaoud; Husrev Taha Sencar; Sanjay Chawla Building Resilience to Out-of-Distribution Visual Data via Input Optimization and Model Finetuning. (1%)Christopher J. Holder; Majid Khonji; Jorge Dias; Muhammad Shafique 2022-11-28 Adversarial Artifact Detection in EEG-Based Brain-Computer Interfaces. (99%)Xiaoqing Chen; Dongrui Wu Interpretations Cannot Be Trusted: Stealthy and Effective Adversarial Perturbations against Interpretable Deep Learning. (95%)Eldor Abdukhamidov; Mohammed Abuhamad; Simon S. Woo; Eric Chan-Tin; Tamer Abuhmed Training Time Adversarial Attack Aiming the Vulnerability of Continual Learning. (83%)Gyojin Han; Jaehyun Choi; Hyeong Gwon Hong; Junmo Kim Towards More Robust Interpretation via Local Gradient Alignment. (76%)Sunghwan Joo; Seokhyeon Jeong; Juyeon Heo; Adrian Weller; Taesup Moon Understanding the Impact of Adversarial Robustness on Accuracy Disparity. (31%)Yuzheng Hu; Fan Wu; Hongyang Zhang; Han Zhao How Important are Good Method Names in Neural Code Generation? A Model Robustness Perspective. (13%)Guang Yang; Yu Zhou; Wenhua Yang; Tao Yue; Xiang Chen; Taolue Chen Rethinking the Number of Shots in Robust Model-Agnostic Meta-Learning. (8%)Xiaoyue Duan; Guoliang Kang; Runqi Wang; Shumin Han; Song Xue; Tian Wang; Baochang Zhang Attack on Unfair ToS Clause Detection: A Case Study using Universal Adversarial Triggers. (8%)Shanshan Xu; Irina Broda; Rashid Haddad; Marco Negrini; Matthias Grabmair Gamma-convergence of a nonlocal perimeter arising in adversarial machine learning. (3%)Leon Bungert; Kerrek Stinson CoNAL: Anticipating Outliers with Large Language Models. (1%)Albert Xu; Xiang Ren; Robin Jia Learning Antidote Data to Individual Unfairness. (1%)Peizhao Li; Ethan Xia; Hongfu Liu 2022-11-27 Imperceptible Adversarial Attack via Invertible Neural Networks. (99%)Zihan Chen; Ziyue Wang; Junjie Huang; Wentao Zhao; Xiao Liu; Dejian Guan Foiling Explanations in Deep Neural Networks. (98%)Snir Vitrack Tamam; Raz Lapid; Moshe Sipper Navigation as the Attacker Wishes? Towards Building Byzantine-Robust Embodied Agents under Federated Learning. (84%)Yunchao Zhang; Zonglin Di; Kaiwen Zhou; Cihang Xie; Xin Wang Traditional Classification Neural Networks are Good Generators: They are Competitive with DDPMs and GANs. (50%)Guangrun Wang; Philip H. S. Torr Federated Learning Attacks and Defenses: A Survey. (47%)Yao Chen; Yijie Gui; Hong Lin; Wensheng Gan; Yongdong Wu Adversarial Rademacher Complexity of Deep Neural Networks. (47%)Jiancong Xiao; Yanbo Fan; Ruoyu Sun; Zhi-Quan Luo 2022-11-26 Game Theoretic Mixed Experts for Combinational Adversarial Machine Learning. (99%)Ethan Rathbun; Kaleel Mahmood; Sohaib Ahmad; Caiwen Ding; Dijk Marten van 2022-11-25 Boundary Adversarial Examples Against Adversarial Overfitting. (99%)Muhammad Zaid Hameed; Beat Buesser Supervised Contrastive Prototype Learning: Augmentation Free Robust Neural Network. (98%)Iordanis Fostiropoulos; Laurent Itti Beyond Smoothing: Unsupervised Graph Representation Learning with Edge Heterophily Discriminating. (3%)Yixin Liu; Yizhen Zheng; Daokun Zhang; Vincent CS Lee; Shirui Pan TrustGAN: Training safe and trustworthy deep learning models through generative adversarial networks. (1%)Hélion du Mas des Bourboux 2022-11-24 SAGA: Spectral Adversarial Geometric Attack on 3D Meshes. (98%)Tomer Stolik; Itai Lang; Shai Avidan Tracking Dataset IP Use in Deep Neural Networks. (96%)Seonhye Park; Alsharif Abuadbba; Shuo Wang; Kristen Moore; Yansong Gao; Hyoungshick Kim; Surya Nepal Explainable and Safe Reinforcement Learning for Autonomous Air Mobility. (92%)Lei Wang; Hongyu Yang; Yi Lin; Suwan Yin; Yuankai Wu Neural Network Complexity of Chaos and Turbulence. (41%)Tim Whittaker; Romuald A. Janik; Yaron Oz Seeds Don't Lie: An Adaptive Watermarking Framework for Computer Vision Models. (8%)Jacob Shams; Ben Nassi; Ikuya Morikawa; Toshiya Shimizu; Asaf Shabtai; Yuval Elovici Generative Joint Source-Channel Coding for Semantic Image Transmission. (1%)Ecenaz Erdemir; Tze-Yang Tung; Pier Luigi Dragotti; Deniz Gunduz CycleGANWM: A CycleGAN watermarking method for ownership verification. (1%)Dongdong Lin; Benedetta Tondi; Bin Li; Mauro Barni 2022-11-23 Query Efficient Cross-Dataset Transferable Black-Box Attack on Action Recognition. (99%)Rohit Gupta; Naveed Akhtar; Gaurav Kumar Nayak; Ajmal Mian; Mubarak Shah Adversarial Attacks are a Surprisingly Strong Baseline for Poisoning Few-Shot Meta-Learners. (99%)Elre T. Oldewage; John Bronskill; Richard E. Turner Reliable Robustness Evaluation via Automatically Constructed Attack Ensembles. (76%)Shengcai Liu; Fu Peng; Ke Tang Dual Graphs of Polyhedral Decompositions for the Detection of Adversarial Attacks. (62%)Huma Jamil; Yajing Liu; Christina Cole; Nathaniel Blanchard; Emily J. King; Michael Kirby; Christopher Peterson Privacy-Enhancing Optical Embeddings for Lensless Classification. (11%)Eric Bezzam; Martin Vetterli; Matthieu Simeoni Principled Data-Driven Decision Support for Cyber-Forensic Investigations. (1%)Soodeh Atefi; Sakshyam Panda; Manos Panaousis; Aron Laszka Data Provenance Inference in Machine Learning. (1%)Mingxue Xu; Xiang-Yang Li 2022-11-22 Benchmarking Adversarially Robust Quantum Machine Learning at Scale. (99%)Maxwell T. West; Sarah M. Erfani; Christopher Leckie; Martin Sevior; Lloyd C. L. Hollenberg; Muhammad Usman PointCA: Evaluating the Robustness of 3D Point Cloud Completion Models Against Adversarial Examples. (99%)Shengshan Hu; Junwei Zhang; Wei Liu; Junhui Hou; Minghui Li; Leo Yu Zhang; Hai Jin; Lichao Sun Attacking Image Splicing Detection and Localization Algorithms Using Synthetic Traces. (98%)Shengbang Fang; Matthew C Stamm Backdoor Cleansing with Unlabeled Data. (75%)Lu Pang; Tao Sun; Haibin Ling; Chao Chen Improving Robust Generalization by Direct PAC-Bayesian Bound Minimization. (70%)Zifan Wang; Nan Ding; Tomer Levinboim; Xi Chen; Radu Soricut SoK: Inference Attacks and Defenses in Human-Centered Wireless Sensing. (69%)Wei Sun; Tingjun Chen; Neil Gong 2022-11-21 Boosting the Transferability of Adversarial Attacks with Global Momentum Initialization. (99%)Jiafeng Wang; Zhaoyu Chen; Kaixun Jiang; Dingkang Yang; Lingyi Hong; Yan Wang; Wenqiang Zhang Understanding the Vulnerability of Skeleton-based Human Activity Recognition via Black-box Attack. (99%)Yunfeng Diao; He Wang; Tianjia Shao; Yong-Liang Yang; Kun Zhou; David Hogg Self-Ensemble Protection: Training Checkpoints Are Good Data Protectors. (99%)Sizhe Chen; Geng Yuan; Xinwen Cheng; Yifan Gong; Minghai Qin; Yanzhi Wang; Xiaolin Huang Addressing Mistake Severity in Neural Networks with Semantic Knowledge. (92%)Natalie Abreu; Nathan Vaska; Victoria Helus Efficient Generalization Improvement Guided by Random Weight Perturbation. (68%)Tao Li; Weihao Yan; Zehao Lei; Yingwen Wu; Kun Fang; Ming Yang; Xiaolin Huang CLAWSAT: Towards Both Robust and Accurate Code Models. (56%)Jinghan Jia; Shashank Srikant; Tamara Mitrovska; Chuang Gan; Shiyu Chang; Sijia Liu; Una-May O'Reilly Fairness Increases Adversarial Vulnerability. (54%)Cuong Tran; Keyu Zhu; Ferdinando Fioretto; Henternyck Pascal Van Don't Watch Me: A Spatio-Temporal Trojan Attack on Deep-Reinforcement-Learning-Augment Autonomous Driving. (10%)Yinbo Yu; Jiajia Liu SPIN: Simulated Poisoning and Inversion Network for Federated Learning-Based 6G Vehicular Networks. (8%)Sunder Ali Khowaja; Parus Khuwaja; Kapal Dev; Angelos Antonopoulos A Survey on Backdoor Attack and Defense in Natural Language Processing. (2%)Xuan Sheng; Zhaoyang Han; Piji Li; Xiangmao Chang Understanding and Improving Visual Prompting: A Label-Mapping Perspective. (2%)Aochuan Chen; Yuguang Yao; Pin-Yu Chen; Yihua Zhang; Sijia Liu Multi-Level Knowledge Distillation for Out-of-Distribution Detection in Text. (1%)Qianhui Wu; Huiqiang Jiang; Haonan Yin; Börje F. Karlsson; Chin-Yew Lin Privacy in Practice: Private COVID-19 Detection in X-Ray Images. (1%)Lucas Lange; Maja Schneider; Erhard Rahm A Tale of Frozen Clouds: Quantifying the Impact of Algorithmic Complexity Vulnerabilities in Popular Web Servers. (1%)Masudul Hasan Masud Bhuiyan; Cristian-Alexandru Staicu 2022-11-20 Spectral Adversarial Training for Robust Graph Neural Network. (99%)Jintang Li; Jiaying Peng; Liang Chen; Zibin Zheng; Tingting Liang; Qing Ling Invisible Backdoor Attack with Dynamic Triggers against Person Re-identification. (81%)Wenli Sun; Xinyang Jiang; Shuguang Dou; Dongsheng Li; Duoqian Miao; Cheng Deng; Cairong Zhao Taming Reachability Analysis of DNN-Controlled Systems via Abstraction-Based Training. (47%)Jiaxu Tian; Dapeng Zhi; Si Liu; Peixin Wang; Guy Katz; Min Zhang Adversarial Cheap Talk. (8%)Chris Lu; Timon Willi; Alistair Letcher; Jakob Foerster Deep Composite Face Image Attacks: Generation, Vulnerability and Detection. (2%)Jag Mohan Singh; Raghavendra Ramachandra AI-KD: Adversarial learning and Implicit regularization for self-Knowledge Distillation. (2%)Hyungmin Kim; Sungho Suh; Sunghyun Baek; Daehwan Kim; Daun Jeong; Hansang Cho; Junmo Kim 2022-11-19 Towards Adversarial Robustness of Deep Vision Algorithms. (92%)Hanshu Yan Phonemic Adversarial Attack against Audio Recognition in Real World. (87%)Jiakai Wang; Zhendong Chen; Zixin Yin; Qinghong Yang; Xianglong Liu Towards Robust Dataset Learning. (82%)Yihan Wu; Xinda Li; Florian Kerschbaum; Heng Huang; Hongyang Zhang Let Graph be the Go Board: Gradient-free Node Injection Attack for Graph Neural Networks via Reinforcement Learning. (80%)Mingxuan Ju; Yujie Fan; Chuxu Zhang; Yanfang Ye Exploring validation metrics for offline model-based optimisation with diffusion models. (75%)Christopher Beckham; Alexandre Piche; David Vazquez; Christopher Pal Mask Off: Analytic-based Malware Detection By Transfer Learning and Model Personalization. (9%)Amirmohammad Pasdar; Young Choon Lee; Seok-Hee Hong Investigating the Security of EV Charging Mobile Applications As an Attack Surface. (1%)K. Sarieddine; M. A. Sayed; S. Torabi; R. Atallah; C. Assi 2022-11-18 Adversarial Stimuli: Attacking Brain-Computer Interfaces via Perturbed Sensory Events. (98%)Bibek Upadhayay; Vahid Behzadan Adversarial Detection by Approximation of Ensemble Boundary. (91%)T. Windeatt Leveraging Algorithmic Fairness to Mitigate Blackbox Attribute Inference Attacks. (68%)Jan Aalmoes; Vasisht Duddu; Antoine Boutet Invariant Learning via Diffusion Dreamed Distribution Shifts. (10%)Priyatham Kattakinda; Alexander Levine; Soheil Feizi Intrusion Detection in Internet of Things using Convolutional Neural Networks. (1%)Martin Kodys; Zhi Lu; Kar Wai Fok; Vrizlynn L. L. Thing Improving Robustness of TCM-based Robust Steganography with Variable Robustness. (1%)Jimin Zhang; Xianfeng Zhao; Xiaolei He Provable Defense against Backdoor Policies in Reinforcement Learning. (1%)Shubham Kumar Bharti; Xuezhou Zhang; Adish Singla; Xiaojin Zhu Scaling Up Dataset Distillation to ImageNet-1K with Constant Memory. (1%)Justin Cui; Ruochen Wang; Si Si; Cho-Jui Hsieh 2022-11-17 Diagnostics for Deep Neural Networks with Automated Copy/Paste Attacks. (99%)Stephen Casper; Kaivalya Hariharan; Dylan Hadfield-Menell Towards Good Practices in Evaluating Transfer Adversarial Attacks. (93%)Zhengyu Zhao; Hanwei Zhang; Renjue Li; Ronan Sicre; Laurent Amsaleg; Michael Backes Assessing Neural Network Robustness via Adversarial Pivotal Tuning. (92%)Peter Ebert Christensen; Vésteinn Snæbjarnarson; Andrea Dittadi; Serge Belongie; Sagie Benaim UPTON: Unattributable Authorship Text via Data Poisoning. (86%)Ziyao Wang; Thai Le; Dongwon Lee Generalizable Deepfake Detection with Phase-Based Motion Analysis. (50%)Ekta Prashnani; Michael Goebel; B. S. Manjunath More Effective Centrality-Based Attacks on Weighted Networks. (15%)Balume Mburano; Weisheng Si; Qing Cao; Wei Xing Zheng Potential Auto-driving Threat: Universal Rain-removal Attack. (2%)Jinchegn Hu; Jihao Li; Zhuoran Hou; Jingjing Jiang; Cunjia Liu; Yuanjian Zhang Data-Centric Debugging: mitigating model failures via targeted data collection. (1%)Sahil Singla; Atoosa Malemir Chegini; Mazda Moayeri; Soheil Feiz A Tale of Two Cities: Data and Configuration Variances in Robust Deep Learning. (1%)Guanqin Zhang; Jiankun Sun; Feng Xu; H. M. N. Dilum Bandara; Shiping Chen; Yulei Sui; Tim Menzies VeriSparse: Training Verified Locally Robust Sparse Neural Networks from Scratch. (1%)Sawinder Kaur; Yi Xiao; Asif Salekin 2022-11-16 T-SEA: Transfer-based Self-Ensemble Attack on Object Detection. (99%)Hao Huang; Ziyan Chen; Huanran Chen; Yongtao Wang; Kevin Zhang Efficiently Finding Adversarial Examples with DNN Preprocessing. (99%)Avriti Chauhan; Mohammad Afzal; Hrishikesh Karmarkar; Yizhak Elboher; Kumar Madhukar; Guy Katz Improving Interpretability via Regularization of Neural Activation Sensitivity. (92%)Ofir Moshe; Gil Fidel; Ron Bitton; Asaf Shabtai Attacking Object Detector Using A Universal Targeted Label-Switch Patch. (86%)Avishag Shapira; Ron Bitton; Dan Avraham; Alon Zolfi; Yuval Elovici; Asaf Shabtai Differentially Private Optimizers Can Learn Adversarially Robust Models. (83%)Yuan Zhang; Zhiqi Bu Interpretable Dimensionality Reduction by Feature Preserving Manifold Approximation and Projection. (56%)Yang Yang; Hongjian Sun; Jialei Gong; Di Yu Privacy against Real-Time Speech Emotion Detection via Acoustic Adversarial Evasion of Machine Learning. (38%)Brian Testa; Yi Xiao; Harshit Sharma; Avery Gump; Asif Salekin Holistic Evaluation of Language Models. (2%)Percy Liang; Rishi Bommasani; Tony Lee; Dimitris Tsipras; Dilara Soylu; Michihiro Yasunaga; Yian Zhang; Deepak Narayanan; Yuhuai Wu; Ananya Kumar; Benjamin Newman; Binhang Yuan; Bobby Yan; Ce Zhang; Christian Cosgrove; Christopher D. Manning; Christopher Ré; Diana Acosta-Navas; Drew A. Hudson; Eric Zelikman; Esin Durmus; Faisal Ladhak; Frieda Rong; Hongyu Ren; Huaxiu Yao; Jue Wang; Keshav Santhanam; Laurel Orr; Lucia Zheng; Mert Yuksekgonul; Mirac Suzgun; Nathan Kim; Neel Guha; Niladri Chatterji; Omar Khattab; Peter Henderson; Qian Huang; Ryan Chi; Sang Michael Xie; Shibani Santurkar; Surya Ganguli; Tatsunori Hashimoto; Thomas Icard; Tianyi Zhang; Vishrav Chaudhary; William Wang; Xuechen Li; Yifan Mai; Yuhui Zhang; Yuta Koreeda Analysis and Detectability of Offline Data Poisoning Attacks on Linear Systems. (1%)Alessio Russo; Alexandre Proutiere 2022-11-15 Resisting Graph Adversarial Attack via Cooperative Homophilous Augmentation. (99%)Zhihao Zhu; Chenwang Wu; Min Zhou; Hao Liao; Defu Lian; Enhong Chen Universal Distributional Decision-based Black-box Adversarial Attack with Reinforcement Learning. (99%)Yiran Huang; Yexu Zhou; Michael Hefenbrock; Till Riedel; Likun Fang; Michael Beigl MORA: Improving Ensemble Robustness Evaluation with Model-Reweighing Attack. (99%)Yunrui Yu; Xitong Gao; Cheng-Zhong Xu Person Text-Image Matching via Text-Featur Interpretability Embedding and External Attack Node Implantation. (92%)Fan Li; Hang Zhou; Huafeng Li; Yafei Zhang; Zhengtao Yu Backdoor Attacks on Time Series: A Generative Approach. (70%)Yujing Jiang; Xingjun Ma; Sarah Monazam Erfani; James Bailey CorruptEncoder: Data Poisoning based Backdoor Attacks to Contrastive Learning. (61%)Jinghuai Zhang; Hongbin Liu; Jinyuan Jia; Neil Zhenqiang Gong Improved techniques for deterministic l2 robustness. (22%)Sahil Singla; Soheil Feizi Backdoor Attacks for Remote Sensing Data with Wavelet Transform. (12%)Nikolaus Dräger; Yonghao Xu; Pedram Ghamisi 2022-11-14 Efficient Adversarial Training with Robust Early-Bird Tickets. (92%)Zhiheng Xi; Rui Zheng; Tao Gui; Qi Zhang; Xuanjing Huang Attacking Face Recognition with T-shirts: Database, Vulnerability Assessment and Detection. (13%)M. Ibsen; C. Rathgeb; F. Brechtel; R. Klepp; K. Pöppelmann; A. George; S. Marcel; C. Busch Towards Robust Numerical Question Answering: Diagnosing Numerical Capabilities of NLP Systems. (5%)Jialiang Xu; Mengyu Zhou; Xinyi He; Shi Han; Dongmei Zhang Explainer Divergence Scores (EDS): Some Post-Hoc Explanations May be Effective for Detecting Unknown Spurious Correlations. (5%)Shea Cardozo; Gabriel Islas Montero; Dmitry Kazhdan; Botty Dimanov; Maleakhi Wijaya; Mateja Jamnik; Pietro Lio Robustifying Deep Vision Models Through Shape Sensitization. (2%)Aditay Tripathi; Rishubh Singh; Anirban Chakraborty; Pradeep Shenoy 2022-11-13 Certifying Robustness of Convolutional Neural Networks with Tight Linear Approximation. (26%)Yuan Xiao; Tongtong Bai; Mingzheng Gu; Chunrong Fang; Zhenyu Chen 2022-11-12 Adversarial and Random Transformations for Robust Domain Adaptation and Generalization. (75%)Liang Xiao; Jiaolong Xu; Dawei Zhao; Erke Shang; Qi Zhu; Bin Dai DriftRec: Adapting diffusion models to blind JPEG restoration. (1%)Simon Welker; Henry N. Chapman; Timo Gerkmann 2022-11-11 Generating Textual Adversaries with Minimal Perturbation. (98%)Xingyi Zhao; Lu Zhang; Depeng Xu; Shuhan Yuan On the robustness of non-intrusive speech quality model by adversarial examples. (98%)Hsin-Yi Lin; Huan-Hsin Tseng; Yu Tsao An investigation of security controls and MITRE ATT\&CK techniques. (47%)Md Rayhanur Rahman; Laurie Williams Investigating co-occurrences of MITRE ATT\&CK Techniques. (12%)Md Rayhanur Rahman; Laurie Williams Remapped Cache Layout: Thwarting Cache-Based Side-Channel Attacks with a Hardware Defense. (9%)Wei Song; Rui Hou; Peng Liu; Xiaoxin Li; Peinan Li; Lutan Zhao; Xiaofei Fu; Yifei Sun; Dan Meng 2022-11-10 Test-time adversarial detection and robustness for localizing humans using ultra wide band channel impulse responses. (99%)Abhiram Kolli; Muhammad Jehanzeb Mirza; Horst Possegger; Horst Bischof Impact of Adversarial Training on Robustness and Generalizability of Language Models. (99%)Enes Altinisik; Hassan Sajjad; Husrev Taha Sencar; Safa Messaoud; Sanjay Chawla Privacy-Utility Balanced Voice De-Identification Using Adversarial Examples. (98%)Meng Chen; Li Lu; Jiadi Yu; Yingying Chen; Zhongjie Ba; Feng Lin; Kui Ren Stay Home Safe with Starving Federated Data. (80%)Jaechul Roh; Yajun Fang MSDT: Masked Language Model Scoring Defense in Text Domain. (38%)Jaechul Roh; Minhao Cheng; Yajun Fang Robust DNN Surrogate Models with Uncertainty Quantification via Adversarial Training. (3%)Lixiang Zhang; Jia Li Mitigating Forgetting in Online Continual Learning via Contrasting Semantically Distinct Augmentations. (1%)Sheng-Feng Yu; Wei-Chen Chiu 2022-11-09 On the Robustness of Explanations of Deep Neural Network Models: A Survey. (50%)Amlan Jyoti; Karthik Balaji Ganesh; Manoj Gayala; Nandita Lakshmi Tunuguntla; Sandesh Kamath; Vineeth N Balasubramanian Are All Edges Necessary? A Unified Framework for Graph Purification. (5%)Zishan Gu; Jintang Li; Liang Chen QuerySnout: Automating the Discovery of Attribute Inference Attacks against Query-Based Systems. (3%)Ana-Maria Cretu; Florimond Houssiau; Antoine Cully; Montjoye Yves-Alexandre de Accountable and Explainable Methods for Complex Reasoning over Text. (2%)Pepa Atanasova Directional Privacy for Deep Learning. (1%)Pedro Faustini; Natasha Fernandes; Shakila Tonni; Annabelle McIver; Mark Dras 2022-11-08 Preserving Semantics in Textual Adversarial Attacks. (99%)David Herel; Hugo Cisneros; Tomas Mikolov NaturalAdversaries: Can Naturalistic Adversaries Be as Effective as Artificial Adversaries? (98%)Saadia Gabriel; Hamid Palangi; Yejin Choi How Fraudster Detection Contributes to Robust Recommendation. (67%)Yuni Lai; Kai Zhou Lipschitz Continuous Algorithms for Graph Problems. (16%)Soh Kumabe; Yuichi Yoshida Learning advisor networks for noisy image classification. (1%)Simone Ricci; Tiberio Uricchio; Bimbo Alberto Del 2022-11-07 Are AlphaZero-like Agents Robust to Adversarial Perturbations? (99%)Li-Cheng Lan; Huan Zhang; Ti-Rong Wu; Meng-Yu Tsai; I-Chen Wu; Cho-Jui Hsieh Black-Box Attack against GAN-Generated Image Detector with Contrastive Perturbation. (82%)Zijie Lou; Gang Cao; Man Lin Deviations in Representations Induced by Adversarial Attacks. (70%)Daniel Steinberg; Paul Munro A Hypergraph-Based Machine Learning Ensemble Network Intrusion Detection System. (1%)Zong-Zhi Lin; Thomas D. Pike; Mark M. Bailey; Nathaniel D. Bastian Interpreting deep learning output for out-of-distribution detection. (1%)Damian Matuszewski; Ida-Maria Sintorn Resilience of Wireless Ad Hoc Federated Learning against Model Poisoning Attacks. (1%)Naoya Tezuka; Hideya Ochiai; Yuwei Sun; Hiroshi Esaki 2022-11-06 Contrastive Weighted Learning for Near-Infrared Gaze Estimation. (31%)Adam Lee 2022-11-05 Textual Manifold-based Defense Against Natural Language Adversarial Examples. (99%)Dang Minh Nguyen; Luu Anh Tuan Stateful Detection of Adversarial Reprogramming. (96%)Yang Zheng; Xiaoyi Feng; Zhaoqiang Xia; Xiaoyue Jiang; Maura Pintor; Ambra Demontis; Battista Biggio; Fabio Roli Robust Lottery Tickets for Pre-trained Language Models. (83%)Rui Zheng; Rong Bao; Yuhao Zhou; Di Liang; Sirui Wang; Wei Wu; Tao Gui; Qi Zhang; Xuanjing Huang 2022-11-04 Improving Adversarial Robustness to Sensitivity and Invariance Attacks with Deep Metric Learning. (99%)Anaelia Ovalle; Evan Czyzycki; Cho-Jui Hsieh Logits are predictive of network type. (68%)Ali Borji An Adversarial Robustness Perspective on the Topology of Neural Networks. (64%)Morgane Goibert; Thomas Ricatte; Elvis Dohmatob Fairness-aware Regression Robust to Adversarial Attacks. (38%)Yulu Jin; Lifeng Lai Extension of Simple Algorithms to the Matroid Secretary Problem. (9%)Simon Park Robustness of Fusion-based Multimodal Classifiers to Cross-Modal Content Dilutions. (3%)Gaurav Verma; Vishwa Vinay; Ryan A. Rossi; Srijan Kumar Data Models for Dataset Drift Controls in Machine Learning With Images. (1%)Luis Oala; Marco Aversa; Gabriel Nobis; Kurt Willis; Yoan Neuenschwander; Michèle Buck; Christian Matek; Jerome Extermann; Enrico Pomarico; Wojciech Samek; Roderick Murray-Smith; Christoph Clausen; Bruno Sanguinetti 2022-11-03 Physically Adversarial Attacks and Defenses in Computer Vision: A Survey. (99%)Xingxing Wei; Bangzheng Pu; Jiefan Lu; Baoyuan Wu Adversarial Defense via Neural Oscillation inspired Gradient Masking. (98%)Chunming Jiang; Yilei Zhang M-to-N Backdoor Paradigm: A Stealthy and Fuzzy Attack to Deep Learning Models. (98%)Linshan Hou; Zhongyun Hua; Yuhong Li; Leo Yu Zhang Robust Few-shot Learning Without Using any Adversarial Samples. (89%)Gaurav Kumar Nayak; Ruchit Rawal; Inder Khatri; Anirban Chakraborty Data-free Defense of Black Box Models Against Adversarial Attacks. (84%)Gaurav Kumar Nayak; Inder Khatri; Shubham Randive; Ruchit Rawal; Anirban Chakraborty Leveraging Domain Features for Detecting Adversarial Attacks Against Deep Speech Recognition in Noise. (38%)Christian Heider Nielsen; Zheng-Hua Tan Try to Avoid Attacks: A Federated Data Sanitization Defense for Healthcare IoMT Systems. (33%)Chong Chen; Ying Gao; Leyu Shi; Siquan Huang Unintended Memorization and Timing Attacks in Named Entity Recognition Models. (12%)Rana Salal Ali; Benjamin Zi Hao Zhao; Hassan Jameel Asghar; Tham Nguyen; Ian David Wood; Dali Kaafar 2022-11-02 Defending with Errors: Approximate Computing for Robustness of Deep Neural Networks. (99%)Amira Guesmi; Ihsen Alouani; Khaled N. Khasawneh; Mouna Baklouti; Tarek Frikha; Mohamed Abid; Nael Abu-Ghazaleh Improving transferability of 3D adversarial attacks with scale and shear transformations. (99%)Jinali Zhang; Yinpeng Dong; Jun Zhu; Jihong Zhu; Minchi Kuang; Xiaming Yuan Certified Robustness of Quantum Classifiers against Adversarial Examples through Quantum Noise. (99%)Jhih-Cing Huang; Yu-Lin Tsai; Chao-Han Huck Yang; Cheng-Fang Su; Chia-Mu Yu; Pin-Yu Chen; Sy-Yen Kuo Adversarial Attack on Radar-based Environment Perception Systems. (99%)Amira Guesmi; Ihsen Alouani Isometric Representations in Neural Networks Improve Robustness. (62%)Kosio Beshkov; Jonas Verhellen; Mikkel Elle Lepperød BATT: Backdoor Attack with Transformation-based Triggers. (56%)Tong Xu; Yiming Li; Yong Jiang; Shu-Tao Xia Untargeted Backdoor Attack against Object Detection. (50%)Chengxiao Luo; Yiming Li; Yong Jiang; Shu-Tao Xia Generative Adversarial Training Can Improve Neural Language Models. (33%)Sajad Movahedi; Azadeh Shakery Backdoor Defense via Suppressing Model Shortcuts. (3%)Sheng Yang; Yiming Li; Yong Jiang; Shu-Tao Xia Human-in-the-Loop Mixup. (1%)Katherine M. Collins; Umang Bhatt; Weiyang Liu; Vihari Piratla; Ilia Sucholutsky; Bradley Love; Adrian Weller 2022-11-01 The Enemy of My Enemy is My Friend: Exploring Inverse Adversaries for Improving Adversarial Training. (99%)Junhao Dong; Seyed-Mohsen Moosavi-Dezfooli; Jianhuang Lai; Xiaohua Xie LMD: A Learnable Mask Network to Detect Adversarial Examples for Speaker Verification. (99%)Xing Chen; Jie Wang; Xiao-Lei Zhang; Wei-Qiang Zhang; Kunde Yang DensePure: Understanding Diffusion Models towards Adversarial Robustness. (98%)Chaowei Xiao; Zhongzhu Chen; Kun Jin; Jiongxiao Wang; Weili Nie; Mingyan Liu; Anima Anandkumar; Bo Li; Dawn Song Adversarial Training with Complementary Labels: On the Benefit of Gradually Informative Attacks. (87%)Jianan Zhou; Jianing Zhu; Jingfeng Zhang; Tongliang Liu; Gang Niu; Bo Han; Masashi Sugiyama Universal Perturbation Attack on Differentiable No-Reference Image- and Video-Quality Metrics. (82%)Ekaterina Shumitskaya; Anastasia Antsiferova; Dmitriy Vatolin The Perils of Learning From Unlabeled Data: Backdoor Attacks on Semi-supervised Learning. (80%)Virat Shejwalkar; Lingjuan Lyu; Amir Houmansadr Maximum Likelihood Distillation for Robust Modulation Classification. (69%)Javier Maroto; Gérôme Bovet; Pascal Frossard FRSUM: Towards Faithful Abstractive Summarization via Enhancing Factual Robustness. (45%)Wenhao Wu; Wei Li; Jiachen Liu; Xinyan Xiao; Ziqiang Cao; Sujian Li; Hua Wu Amplifying Membership Exposure via Data Poisoning. (22%)Yufei Chen; Chao Shen; Yun Shen; Cong Wang; Yang Zhang ActGraph: Prioritization of Test Cases Based on Deep Neural Network Activation Graph. (13%)Jinyin Chen; Jie Ge; Haibin Zheng 2022-10-31 Scoring Black-Box Models for Adversarial Robustness. (98%)Jian Vora; Pranay Reddy Samala ARDIR: Improving Robustness using Knowledge Distillation of Internal Representation. (88%)Tomokatsu Takahashi; Masanori Yamada; Yuuki Yamanaka; Tomoya Yamashita SoK: Modeling Explainability in Security Analytics for Interpretability, Trustworthiness, and Usability. (33%)Dipkamal Bhusal; Rosalyn Shin; Ajay Ashok Shewale; Monish Kumar Manikya Veerabhadran; Michael Clifford; Sara Rampazzi; Nidhi Rastogi Preventing Verbatim Memorization in Language Models Gives a False Sense of Privacy. (16%)Daphne Ippolito; Florian Tramèr; Milad Nasr; Chiyuan Zhang; Matthew Jagielski; Katherine Lee; Christopher A. Choquette-Choo; Nicholas Carlini 2022-10-30 Poison Attack and Defense on Deep Source Code Processing Models. (99%)Jia Li; Zhuo Li; Huangzhao Zhang; Ge Li; Zhi Jin; Xing Hu; Xin Xia Character-level White-Box Adversarial Attacks against Transformers via Attachable Subwords Substitution. (99%)Aiwei Liu; Honghai Yu; Xuming Hu; Shu'ang Li; Li Lin; Fukun Ma; Yawen Yang; Lijie Wen Benchmarking Adversarial Patch Against Aerial Detection. (99%)Jiawei Lian; Shaohui Mei; Shun Zhang; Mingyang Ma Symmetric Saliency-based Adversarial Attack To Speaker Identification. (92%)Jiadi Yao; Xing Chen; Xiao-Lei Zhang; Wei-Qiang Zhang; Kunde Yang FI-ODE: Certified and Robust Forward Invariance in Neural ODEs. (61%)Yujia Huang; Ivan Dario Jimenez Rodriguez; Huan Zhang; Yuanyuan Shi; Yisong Yue Imitating Opponent to Win: Adversarial Policy Imitation Learning in Two-player Competitive Games. (9%)The Viet Bui; Tien Mai; Thanh H. Nguyen 2022-10-29 On the Need of Neuromorphic Twins to Detect Denial-of-Service Attacks on Communication Networks. (10%)Holger Boche; Rafael F. Schaefer; H. Vincent Poor; Frank H. P. Fitzek 2022-10-28 Universal Adversarial Directions. (99%)Ching Lam Choi; Farzan Farnia Improving the Transferability of Adversarial Attacks on Face Recognition with Beneficial Perturbation Feature Augmentation. (99%)Fengfan Zhou; Hefei Ling; Yuxuan Shi; Jiazhong Chen; Zongyi Li; Ping Li Improving Hyperspectral Adversarial Robustness Under Multiple Attacks. (98%)Nicholas Soucy; Salimeh Yasaei Sekeh Distributed Black-box Attack against Image Classification Cloud Services. (95%)Han Wu; Sareh Rowlands; Johan Wahlstrom RoChBert: Towards Robust BERT Fine-tuning for Chinese. (75%)Zihan Zhang; Jinfeng Li; Ning Shi; Bo Yuan; Xiangyu Liu; Rong Zhang; Hui Xue; Donghong Sun; Chao Zhang Robust Boosting Forests with Richer Deep Feature Hierarchy. (56%)Jianqiao Wangni Localized Randomized Smoothing for Collective Robustness Certification. (26%)Jan Schuchardt; Tom Wollschläger; Aleksandar Bojchevski; Stephan Günnemann Towards Reliable Neural Specifications. (11%)Chuqin Geng; Nham Le; Xiaojie Xu; Zhaoyue Wang; Arie Gurfinkel; Xujie Si On the Vulnerability of Data Points under Multiple Membership Inference Attacks and Target Models. (1%)Mauro Conti; Jiaxin Li; Stjepan Picek 2022-10-27 TAD: Transfer Learning-based Multi-Adversarial Detection of Evasion Attacks against Network Intrusion Detection Systems. (99%)Islam Debicha; Richard Bauwens; Thibault Debatty; Jean-Michel Dricot; Tayeb Kenaza; Wim Mees Isometric 3D Adversarial Examples in the Physical World. (99%)Yibo Miao; Yinpeng Dong; Jun Zhu; Xiao-Shan Gao LeNo: Adversarial Robust Salient Object Detection Networks with Learnable Noise. (92%)He Tang; He Wang TASA: Deceiving Question Answering Models by Twin Answer Sentences Attack. (92%)Yu Cao; Dianqi Li; Meng Fang; Tianyi Zhou; Jun Gao; Yibing Zhan; Dacheng Tao Efficient and Effective Augmentation Strategy for Adversarial Training. (56%)Sravanti Addepalli; Samyak Jain; R. Venkatesh Babu Noise Injection Node Regularization for Robust Learning. (2%)Noam Levi; Itay M. Bloch; Marat Freytsis; Tomer Volansky Domain Adaptive Object Detection for Autonomous Driving under Foggy Weather. (1%)Jinlong Li; Runsheng Xu; Jin Ma; Qin Zou; Jiaqi Ma; Hongkai Yu 2022-10-26 Improving Adversarial Robustness with Self-Paced Hard-Class Pair Reweighting. (99%)Pengyue Hou; Jie Han; Xingyu Li There is more than one kind of robustness: Fooling Whisper with adversarial examples. (98%)Raphael Olivier; Bhiksha Raj Disentangled Text Representation Learning with Information-Theoretic Perspective for Adversarial Robustness. (86%)Jiahao Zhao; Wenji Mao BioNLI: Generating a Biomedical NLI Dataset Using Lexico-semantic Constraints for Adversarial Examples. (75%)Mohaddeseh Bastan; Mihai Surdeanu; Niranjan Balasubramanian EIPSIM: Modeling Secure IP Address Allocation at Cloud Scale. (11%)Eric University of Wisconsin-Madison Pauley; Kyle Pennsylvania State University Domico; Blaine University of Wisconsin-Madison Hoak; Ryan University of Wisconsin-Madison Sheatsley; Quinn University of Wisconsin-Madison Burke; Yohan University of Wisconsin-Madison Beugin; Patrick University of Wisconsin-Madison McDaniel V-Cloak: Intelligibility-, Naturalness- & Timbre-Preserving Real-Time Voice Anonymization. (10%)Jiangyi Zhejiang University Deng; Fei Zhejiang University Teng; Yanjiao Zhejiang University Chen; Xiaofu Wuhan University Chen; Zhaohui Wuhan University Wang; Wenyuan Zhejiang University Xu Rethinking the Reverse-engineering of Trojan Triggers. (5%)Zhenting Wang; Kai Mei; Hailun Ding; Juan Zhai; Shiqing Ma Cover Reproducible Steganography via Deep Generative Models. (1%)Kejiang Chen; Hang Zhou; Yaofei Wang; Menghan Li; Weiming Zhang; Nenghai Yu DEMIS: A Threat Model for Selectively Encrypted Visual Surveillance Data. (1%)Ifeoluwapo Aribilola; Mamoona Naveed Asghar; Brian Lee Privately Fine-Tuning Large Language Models with Differential Privacy. (1%)Rouzbeh Behnia; Mohamamdreza Ebrahimi; Jason Pacheco; Balaji Padmanabhan 2022-10-25 LP-BFGS attack: An adversarial attack based on the Hessian with limited pixels. (99%)Jiebao Zhang; Wenhua Qian; Rencan Nie; Jinde Cao; Dan Xu Adversarially Robust Medical Classification via Attentive Convolutional Neural Networks. (99%)Isaac Wasserman A White-Box Adversarial Attack Against a Digital Twin. (99%)Wilson Patterson; Ivan Fernandez; Subash Neupane; Milan Parmar; Sudip Mittal; Shahram Rahimi Multi-view Representation Learning from Malware to Defend Against Adversarial Variants. (98%)James Lee Hu; Mohammadreza Ebrahimi; Weifeng Li; Xin Li; Hsinchun Chen Adversarial Purification with the Manifold Hypothesis. (98%)Zhaoyuan Yang; Zhiwei Xu; Jing Zhang; Richard Hartley; Peter Tu Improving Adversarial Robustness via Joint Classification and Multiple Explicit Detection Classes. (98%)Sina Baharlouei; Fatemeh Sheikholeslami; Meisam Razaviyayn; Zico Kolter Accelerating Certified Robustness Training via Knowledge Transfer. (73%)Pratik Vaishnavi; Kevin Eykholt; Amir Rahmati Causal Information Bottleneck Boosts Adversarial Robustness of Deep Neural Network. (64%)Huan Hua; Jun Yan; Xi Fang; Weiquan Huang; Huilin Yin; Wancheng Ge Towards Robust Recommender Systems via Triple Cooperative Defense. (61%)Qingyang Wang; Defu Lian; Chenwang Wu; Enhong Chen Towards Formal Approximated Minimal Explanations of Neural Networks. (13%)Shahaf Bassan; Guy Katz FocusedCleaner: Sanitizing Poisoned Graphs for Robust GNN-based Node Classification. (13%)Yulin Zhu; Liang Tong; Kai Zhou A Streamlit-based Artificial Intelligence Trust Platform for Next-Generation Wireless Networks. (3%)M. Kuzlu; F. O. Catak; S. Sarp; U. Cali; O Gueler Robustness of Locally Differentially Private Graph Analysis Against Poisoning. (1%)Jacob Imola; Amrita Roy Chowdhury; Kamalika Chaudhuri 2022-10-24 Ares: A System-Oriented Wargame Framework for Adversarial ML. (99%)Farhan Ahmed; Pratik Vaishnavi; Kevin Eykholt; Amir Rahmati SpacePhish: The Evasion-space of Adversarial Attacks against Phishing Website Detectors using Machine Learning. (99%)Giovanni Apruzzese; Mauro Conti; Ying Yuan Motif-Backdoor: Rethinking the Backdoor Attack on Graph Neural Networks via Motifs. (96%)Haibin Zheng; Haiyang Xiong; Jinyin Chen; Haonan Ma; Guohan Huang On the Robustness of Dataset Inference. (88%)Sebastian Szyller; Rui Zhang; Jian Liu; N. Asokan Flexible Android Malware Detection Model based on Generative Adversarial Networks with Code Tensor. (16%)Zhao Yang; Fengyang Deng; Linxi Han Revisiting Sparse Convolutional Model for Visual Recognition. (11%)Xili Dai; Mingyang Li; Pengyuan Zhai; Shengbang Tong; Xingjian Gao; Shao-Lun Huang; Zhihui Zhu; Chong You; Yi Ma 2022-10-23 FLIP: A Provable Defense Framework for Backdoor Mitigation in Federated Learning. (68%)Kaiyuan Zhang; Guanhong Tao; Qiuling Xu; Siyuan Cheng; Shengwei An; Yingqi Liu; Shiwei Feng; Guangyu Shen; Pin-Yu Chen; Shiqing Ma; Xiangyu Zhang Adversarial Pretraining of Self-Supervised Deep Networks: Past, Present and Future. (45%)Guo-Jun Qi; Mubarak Shah 2022-10-22 ADDMU: Detection of Far-Boundary Adversarial Examples with Data and Model Uncertainty Estimation. (99%)Fan Yin; Yao Li; Cho-Jui Hsieh; Kai-Wei Chang Hindering Adversarial Attacks with Implicit Neural Representations. (92%)Andrei A. Rusu; Dan A. Calian; Sven Gowal; Raia Hadsell GANI: Global Attacks on Graph Neural Networks via Imperceptible Node Injections. (81%)Junyuan Fang; Haixian Wen; Jiajing Wu; Qi Xuan; Zibin Zheng; Chi K. Tse Nash Equilibria and Pitfalls of Adversarial Training in Adversarial Robustness Games. (26%)Maria-Florina Balcan; Rattana Pukdee; Pradeep Ravikumar; Hongyang Zhang Precisely the Point: Adversarial Augmentations for Faithful and Informative Text Generation. (4%)Wenhao Wu; Wei Li; Jiachen Liu; Xinyan Xiao; Sujian Li; Yajuan Lyu 2022-10-21 Evolution of Neural Tangent Kernels under Benign and Adversarial Training. (99%)Noel Loo; Ramin Hasani; Alexander Amini; Daniela Rus The Dark Side of AutoML: Towards Architectural Backdoor Search. (68%)Ren Pang; Changjiang Li; Zhaohan Xi; Shouling Ji; Ting Wang Diffusion Visual Counterfactual Explanations. (10%)Maximilian Augustin; Valentyn Boreiko; Francesco Croce; Matthias Hein TCAB: A Large-Scale Text Classification Attack Benchmark. (10%)Kalyani Asthana; Zhouhang Xie; Wencong You; Adam Noack; Jonathan Brophy; Sameer Singh; Daniel Lowd A critical review of cyber-physical security for building automation systems. (2%)Guowen Li; Lingyu Ren; Yangyang Fu; Zhiyao Yang; Veronica Adetola; Jin Wen; Qi Zhu; Teresa Wu; K. Selcuk Candanf; Zheng O'Neill Extracted BERT Model Leaks More Information than You Think! (1%)Xuanli He; Chen Chen; Lingjuan Lyu; Qiongkai Xu 2022-10-20 Identifying Human Strategies for Generating Word-Level Adversarial Examples. (98%)Maximilian Mozes; Bennett Kleinberg; Lewis D. Griffin Are You Stealing My Model? Sample Correlation for Fingerprinting Deep Neural Networks. (98%)Jiyang Guan; Jian Liang; Ran He Balanced Adversarial Training: Balancing Tradeoffs between Fickleness and Obstinacy in NLP Models. (98%)Hannah Chen; Yangfeng Ji; David Evans Learning Sample Reweighting for Accuracy and Adversarial Robustness. (93%)Chester Holtz; Tsui-Wei Weng; Gal Mishne Similarity of Neural Architectures using Adversarial Attack Transferability. (86%)Jaehui Hwang; Dongyoon Han; Byeongho Heo; Song Park; Sanghyuk Chun; Jong-Seok Lee New data poison attacks on machine learning classifiers for mobile exfiltration. (80%)Miguel A. Ramirez; Sangyoung Yoon; Ernesto Damiani; Hussam Al Hamadi; Claudio Agostino Ardagna; Nicola Bena; Young-Ji Byon; Tae-Yeon Kim; Chung-Suk Cho; Chan Yeob Yeun Attacking Motion Estimation with Adversarial Snow. (16%)Jenny Schmalfuss; Lukas Mehl; Andrés Bruhn How Does a Deep Learning Model Architecture Impact Its Privacy? A Comprehensive Study of Privacy Attacks on CNNs and Transformers. (13%)Guangsheng Zhang; Bo Liu; Huan Tian; Tianqing Zhu; Ming Ding; Wanlei Zhou Analyzing the Robustness of Decentralized Horizontal and Vertical Federated Learning Architectures in a Non-IID Scenario. (4%)Pedro Miguel Sánchez Sánchez; Alberto Huertas Celdrán; Enrique Tomás Martínez Beltrán; Daniel Demeter; Gérôme Bovet; Gregorio Martínez Pérez; Burkhard Stiller Apple of Sodom: Hidden Backdoors in Superior Sentence Embeddings via Contrastive Learning. (3%)Xiaoyi Chen; Baisong Xin; Shengfang Zhai; Shiqing Ma; Qingni Shen; Zhonghai Wu LOT: Layer-wise Orthogonal Training on Improving $\ell_2$ Certified Robustness. (3%)Xiaojun Xu; Linyi Li; Bo Li 2022-10-19 Learning Transferable Adversarial Robust Representations via Multi-view Consistency. (99%)Minseon Kim; Hyeonjeong Ha; Dong Bok Lee; Sung Ju Hwang Effective Targeted Attacks for Adversarial Self-Supervised Learning. (99%)Minseon Kim; Hyeonjeong Ha; Sooel Son; Sung Ju Hwang No-Box Attacks on 3D Point Cloud Classification. (93%)Hanieh Naderi; Chinthaka Dinesh; Ivan V. Bajic; Shohreh Kasaei Backdoor Attack and Defense in Federated Generative Adversarial Network-based Medical Image Synthesis. (83%)Ruinan Jin; Xiaoxiao Li Chaos Theory and Adversarial Robustness. (73%)Jonathan S. Kent Emerging Threats in Deep Learning-Based Autonomous Driving: A Comprehensive Survey. (69%)Hui Cao; Wenlong Zou; Yinkun Wang; Ting Song; Mengjun Liu Why Should Adversarial Perturbations be Imperceptible? Rethink the Research Paradigm in Adversarial NLP. (64%)Yangyi Chen; Hongcheng Gao; Ganqu Cui; Fanchao Qi; Longtao Huang; Zhiyuan Liu; Maosong Sun FedRecover: Recovering from Poisoning Attacks in Federated Learning using Historical Information. (41%)Xiaoyu Cao; Jinyuan Jia; Zaixi Zhang; Neil Zhenqiang Gong Learning to Invert: Simple Adaptive Attacks for Gradient Inversion in Federated Learning. (16%)Ruihan Wu; Xiangyu Chen; Chuan Guo; Kilian Q. Weinberger Variational Model Perturbation for Source-Free Domain Adaptation. (1%)Mengmeng Jing; Xiantong Zhen; Jingjing Li; Cees G. M. Snoek 2022-10-18 Scaling Adversarial Training to Large Perturbation Bounds. (98%)Sravanti Addepalli; Samyak Jain; Gaurang Sriramanan; R. Venkatesh Babu Not All Poisons are Created Equal: Robust Training against Data Poisoning. (97%)Yu Yang; Tian Yu Liu; Baharan Mirzasoleiman ROSE: Robust Selective Fine-tuning for Pre-trained Language Models. (73%)Lan Jiang; Hao Zhou; Yankai Lin; Peng Li; Jie Zhou; Rui Jiang Analysis of Master Vein Attacks on Finger Vein Recognition Systems. (56%)Huy H. Nguyen; Trung-Nghia Le; Junichi Yamagishi; Isao Echizen Training set cleansing of backdoor poisoning by self-supervised representation learning. (56%)H. Wang; S. Karami; O. Dia; H. Ritter; E. Emamjomeh-Zadeh; J. Chen; Z. Xiang; D. J. Miller; G. Kesidis On the Adversarial Robustness of Mixture of Experts. (13%)Joan Puigcerver; Rodolphe Jenatton; Carlos Riquelme; Pranjal Awasthi; Srinadh Bhojanapalli