AIセキュリティポータルbot

Automated and Explainable Denial of Service Analysis for AI-Driven Intrusion Detection Systems

Authors: Paul Badu Yakubu, Lesther Santana, Mohamed Rahouti, Yufeng Xin, Abdellah Chehri, Mohammed Aledhari | Published: 2025-11-06
Traffic Characteristic Analysis
Model DoS
Feature Importance Analysis

Specification-Guided Vulnerability Detection with Large Language Models

Authors: Hao Zhu, Jia Li, Cuiyun Gao, Jiaru Qian, Yihong Dong, Huanyu Liu, Lecheng Wang, Ziliang Wang, Xiaolong Hu, Ge Li | Published: 2025-11-06
Prompt Injection
Large Language Model
脆弱性検出手法

Hybrid Fuzzing with LLM-Guided Input Mutation and Semantic Feedback

Authors: Shiyin Lin | Published: 2025-11-06
Prompt Injection
Dynamic Analysis
Information Security

Whisper Leak: a side-channel attack on Large Language Models

Authors: Geoff McDonald, Jonathan Bar Or | Published: 2025-11-05
Traffic Characteristic Analysis
Prompt leaking
Large Language Model

Watermarking Large Language Models in Europe: Interpreting the AI Act in Light of Technology

Authors: Thomas Souverain | Published: 2025-11-05
Digital Watermarking for Generative AI
Generative Model Characteristics
Transparency and Verification

Let the Bees Find the Weak Spots: A Path Planning Perspective on Multi-Turn Jailbreak Attacks against LLMs

Authors: Yize Liu, Yunyun Hou, Aina Sui | Published: 2025-11-05
Automation of Cybersecurity
Prompt Injection
マルチターン攻撃分析

Auditing M-LLMs for Privacy Risks: A Synthetic Benchmark and Evaluation Framework

Authors: Junhao Li, Jiahao Chen, Zhou Feng, Chunyi Zhou | Published: 2025-11-05
Hallucination
Privacy Violation
Privacy Protection

Death by a Thousand Prompts: Open Model Vulnerability Analysis

Authors: Amy Chang, Nicholas Conley, Harish Santhanalakshmi Ganesan, Adam Swanda | Published: 2025-11-05
Disabling Safety Mechanisms of LLM
Indirect Prompt Injection
Threat modeling

Measuring the Security of Mobile LLM Agents under Adversarial Prompts from Untrusted Third-Party Channels

Authors: Chenghao Du, Quanfeng Huang, Tingxuan Tang, Zihao Wang, Adwait Nadkarni, Yue Xiao | Published: 2025-10-31 | Updated: 2025-11-06
Indirect Prompt Injection
Prompt Injection
Information Security

PVMark: Enabling Public Verifiability for LLM Watermarking Schemes

Authors: Haohua Duan, Liyao Xiang, Xin Zhang | Published: 2025-10-30
Model Extraction Attack
公的検証可能性
Watermarking Technology