Literature Database

On the Effectiveness of Membership Inference in Targeted Data Extraction from Large Language Models

Authors: Ali Al Sahili, Ali Chehab, Razane Tajeddine | Published: 2025-12-15
Data Extraction and Analysis
Prompt leaking
評価メトリクス

Security and Detectability Analysis of Unicode Text Watermarking Methods Against Large Language Models

Authors: Malte Hellmeier | Published: 2025-12-15
セキュリティ検証手法
Digital Watermarking for Generative AI
Watermark Evaluation

Cisco Integrated AI Security and Safety Framework Report

Authors: Amy Chang, Tiffany Saade, Sanket Mendapara, Adam Swanda, Ankit Garg | Published: 2025-12-15
Indirect Prompt Injection
サイバーセキュリティの脅威
セキュリティ考慮

CTIGuardian: A Few-Shot Framework for Mitigating Privacy Leakage in Fine-Tuned LLMs

Authors: Shashie Dilhara Batan Arachchige, Benjamin Zi Hao Zhao, Hassan Jameel Asghar, Dinusha Vatsalan, Dali Kaafar | Published: 2025-12-15
Trade-off Analysis
Privacy Protection Method
Prompt leaking

Virtual camera detection: Catching video injection attacks in remote biometric systems

Authors: Daniyar Kurmankhojayev, Andrei Shadrikov, Dmitrii Gordin, Mikhail Shkorin, Danijar Gabdullin, Aigerim Kambetbayeva, Kanat Kuatov | Published: 2025-12-11
VCD技術
User Identification System
Malicious Client

TriDF: Evaluating Perception, Detection, and Hallucination for Interpretable DeepFake Detection

Authors: Jian-Yu Jiang-Lin, Kang-Yang Huang, Ling Zou, Ling Lo, Sheng-Ping Yang, Yu-Wen Tseng, Kun-Hsiang Lin, Chia-Ling Chen, Yu-Ting Ta, Yan-Tsung Wang, Po-Ching Chen, Hongxia Xie, Hong-Han Shuai, Wen-Huang Cheng | Published: 2025-12-11
Detection of Hallucinations
Model DoS
Evaluation Method

LLM-Assisted AHP for Explainable Cyber Range Evaluation

Authors: Vyron Kampourakis, Georgios Kavallieratos, Georgios Spathoulas, Vasileios Gkioulos, Sokratis Katsikas | Published: 2025-12-11
XAI (Explainable AI)
Reliability Assessment
Evaluation Method

From Lab to Reality: A Practical Evaluation of Deep Learning Models and LLMs for Vulnerability Detection

Authors: Chaomeng Lu, Bert Lagaisse | Published: 2025-12-11
Certified Robustness
Calculation of Output Harmfulness
Evaluation Method

When Reject Turns into Accept: Quantifying the Vulnerability of LLM-Based Scientific Reviewers to Indirect Prompt Injection

Authors: Devanshu Sahoo, Manish Prasad, Vasudev Majhi, Jahnvi Singh, Vinay Chamola, Yash Sinha, Murari Mandal, Dhruv Kumar | Published: 2025-12-11
Indirect Prompt Injection
Adversarial Attack Analysis
Evaluation Method

Differential Privacy for Secure Machine Learning in Healthcare IoT-Cloud Systems

Authors: N Mangala, Murtaza Rangwala, S Aishwarya, B Eswara Reddy, Rajkumar Buyya, KR Venugopal, SS Iyengar, LM Patnaik | Published: 2025-12-11
Detection of Poison Data for Backdoor Attacks
Privacy Enhancing Technology
Differential Privacy