Privacy-Aware In-Context Learning for Large Language Models

Authors: Bishnu Bhusal, Manoj Acharya, Ramneet Kaur, Colin Samplawski, Anirban Roy, Adam D. Cobb, Rohit Chadha, Susmit Jha | Published: 2025-09-17

A Multi-Agent LLM Defense Pipeline Against Prompt Injection Attacks

Authors: S M Asif Hossain, Ruksat Khan Shayoni, Mohd Ruhul Ameen, Akif Islam, M. F. Mridha, Jungpil Shin | Published: 2025-09-16 | Updated: 2025-10-01

Sy-FAR: Symmetry-based Fair Adversarial Robustness

Authors: Haneen Najjar, Eyal Ronen, Mahmood Sharif | Published: 2025-09-16

Jailbreaking Large Language Models Through Content Concretization

Authors: Johan Wahréus, Ahmed Hussain, Panos Papadimitratos | Published: 2025-09-16

A Graph-Based Approach to Alert Contextualisation in Security Operations Centres

Authors: Magnus Wiik Eckhoff, Peter Marius Flydal, Siem Peters, Martin Eian, Jonas Halvorsen, Vasileios Mavroeidis, Gudmund Grov | Published: 2025-09-16

EByFTVeS: Efficient Byzantine Fault Tolerant-based Verifiable Secret-sharing in Distributed Privacy-preserving Machine Learning

Authors: Zhen Li, Zijian Zhang, Wenjin Yang, Pengbo Wang, Zhaoqi Wang, Meng Li, Yan Wu, Xuyang Liu, Jing Sun, Liehuang Zhu | Published: 2025-09-16

A Scalable Architecture for Efficient Multi-bit Fully Homomorphic Encryption

Authors: Jiaao Ma, Ceyu Xu, Lisa Wu Wills | Published: 2025-09-16

A Systematic Evaluation of Parameter-Efficient Fine-Tuning Methods for the Security of Code LLMs

Authors: Kiho Lee, Jungkon Kim, Doowon Kim, Hyoungshick Kim | Published: 2025-09-16

Yet Another Watermark for Large Language Models

Authors: Siyuan Bao, Ying Shi, Zhiguang Yang, Hanzhou Wu, Xinpeng Zhang | Published: 2025-09-16

Time-Constrained Intelligent Adversaries for Automation Vulnerability Testing: A Multi-Robot Patrol Case Study

Authors: James C. Ward, Alex Bott, Connor York, Edmund R. Hunt | Published: 2025-09-15