Literature Database

SALAD: Systematic Assessment of Machine Unlearning on LLM-Aided Hardware Design

Authors: Zeng Wang, Minghao Shao, Rupesh Karn, Likhitha Mankali, Jitendra Bhandari, Ramesh Karri, Ozgur Sinanoglu, Muhammad Shafique, Johann Knechtel | Published: 2025-06-02 | Updated: 2025-08-05
Data-Driven Vulnerability Assessment
Prompt leaking
Watermark

On the Stability of Graph Convolutional Neural Networks: A Probabilistic Perspective

Authors: Ning Zhang, Henry Kenlay, Li Zhang, Mihai Cucuringu, Xiaowen Dong | Published: 2025-06-01 | Updated: 2025-06-03
Dynamic Graph Processing
Adversarial Learning
Optimization Problem

A Large Language Model-Supported Threat Modeling Framework for Transportation Cyber-Physical Systems

Authors: M Sabbir Salek, Mashrur Chowdhury, Muhaimin Bin Munir, Yuchen Cai, Mohammad Imtiaz Hasan, Jean-Michel Tine, Latifur Khan, Mizanur Rahman | Published: 2025-06-01 | Updated: 2025-07-28
Bias Detection in AI Output
Poisoning attack on RAG
Vulnerability Assessment Method

Beyond the Protocol: Unveiling Attack Vectors in the Model Context Protocol (MCP) Ecosystem

Authors: Hao Song, Yiming Shen, Wenxuan Luo, Leixin Guo, Ting Chen, Jiashui Wang, Beibei Li, Xiaosong Zhang, Jiachi Chen | Published: 2025-05-31 | Updated: 2025-08-20
Indirect Prompt Injection
Prompt Injection
Attack Type

A Systematic Review of Metaheuristics-Based and Machine Learning-Driven Intrusion Detection Systems in IoT

Authors: Mohammad Shamim Ahsan, Salekul Islam, Swakkhar Shatabda | Published: 2025-05-31 | Updated: 2025-06-03
Prompt Injection
Intrusion Detection System
Selection and Evaluation of Optimization Algorithms

A Red Teaming Roadmap Towards System-Level Safety

Authors: Zifan Wang, Christina Q. Knight, Jeremy Kritz, Willow E. Primack, Julian Michael | Published: 2025-05-30 | Updated: 2025-06-09
Model DoS
Large Language Model
製品安全性

Rethinking Exact Unlearning under Exposure: Extracting Forgotten Data under Exact Unlearning in Large Language Model

Authors: Xiaoyu Wu, Yifei Pang, Terrance Liu, Zhiwei Steven Wu | Published: 2025-05-30 | Updated: 2025-10-06
Privacy-Preserving Machine Learning
Privacy Loss Analysis
倫理基準遵守

MCP Safety Training: Learning to Refuse Falsely Benign MCP Exploits using Improved Preference Alignment

Authors: John Halloran | Published: 2025-05-29
Poisoning attack on RAG
Alignment
料理材料

Merge Hijacking: Backdoor Attacks to Model Merging of Large Language Models

Authors: Zenghui Yuan, Yangming Xu, Jiawen Shi, Pan Zhou, Lichao Sun | Published: 2025-05-29
LLM Security
Poisoning Attack
Model Protection Methods

SafeCOMM: A Study on Safety Degradation in Fine-Tuned Telecom Large Language Models

Authors: Aladin Djuhera, Swanand Ravindra Kadhe, Farhan Ahmed, Syed Zawad, Fernando Koch, Walid Saad, Holger Boche | Published: 2025-05-29 | Updated: 2025-10-27
Prompt Injection
Large Language Model
安全性評価