Literature Database

Misusing Tools in Large Language Models With Visual Adversarial Examples

Authors: Xiaohan Fu, Zihan Wang, Shuheng Li, Rajesh K. Gupta, Niloofar Mireshghallah, Taylor Berg-Kirkpatrick, Earlence Fernandes | Published: 2023-10-04
LLM Performance Evaluation
Prompt Injection
Adversarial Example

Raze to the Ground: Query-Efficient Adversarial HTML Attacks on Machine-Learning Phishing Webpage Detectors

Authors: Biagio Montaruli, Luca Demetrio, Maura Pintor, Luca Compagna, Davide Balzarotti, Battista Biggio | Published: 2023-10-04 | Updated: 2023-10-14
Phishing
Poisoning
Machine Learning Method

Crossed-IoT device portability of Electromagnetic Side Channel Analysis: Challenges and Dataset

Authors: Tharindu Lakshan Yasarathna, Lojenaa Navanesan, Simon Barque, Assanka Sayakkara, Nhien-An Le-Khac | Published: 2023-10-04
Watermarking
Cross-Device Analysis
Classification Pattern Analysis

Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models

Authors: Xianjun Yang, Xiao Wang, Qi Zhang, Linda Petzold, William Yang Wang, Xun Zhao, Dahua Lin | Published: 2023-10-04
Prompt Injection
Safety Alignment
Malicious Content Generation

AGIR: Automating Cyber Threat Intelligence Reporting with Natural Language Generation

Authors: Filippo Perrina, Francesco Marchiori, Mauro Conti, Nino Vincenzo Verde | Published: 2023-10-04
Data Generation
Threat modeling
evaluation metrics

Practical, Private Assurance of the Value of Collaboration via Fully Homomorphic Encryption

Authors: Hassan Jameel Asghar, Zhigang Lu, Zhongrui Zhao, Dali Kaafar | Published: 2023-10-04 | Updated: 2024-10-07
Watermarking
Privacy Protection
Loss Function

zkFL: Zero-Knowledge Proof-based Gradient Aggregation for Federated Learning

Authors: Zhipeng Wang, Nanqing Dong, Jiahao Sun, William Knottenbelt, Yike Guo | Published: 2023-10-04 | Updated: 2024-05-10
Watermarking
Privacy Protection
Federated Learning

Low-Resource Languages Jailbreak GPT-4

Authors: Zheng-Xin Yong, Cristina Menghini, Stephen H. Bach | Published: 2023-10-03 | Updated: 2024-01-27
Prompt Injection
Safety Alignment
Vulnerability detection

Jailbreaker in Jail: Moving Target Defense for Large Language Models

Authors: Bocheng Chen, Advait Paliwal, Qiben Yan | Published: 2023-10-03
LLM Performance Evaluation
Prompt Injection
evaluation metrics

FLEDGE: Ledger-based Federated Learning Resilient to Inference and Backdoor Attacks

Authors: Jorge Castillo, Phillip Rieger, Hossein Fereidooni, Qian Chen, Ahmad Sadeghi | Published: 2023-10-03
Backdoor Attack
Privacy Protection
Poisoning