LLMセキュリティ

Humanizing Machine-Generated Content: Evading AI-Text Detection through Adversarial Attack

Authors: Ying Zhou, Ben He, Le Sun | Published: 2024-04-02
LLMセキュリティ
ウォーターマーキング
攻撃手法

Can LLMs get help from other LLMs without revealing private information?

Authors: Florian Hartmann, Duc-Hieu Tran, Peter Kairouz, Victor Cărbune, Blaise Aguera y Arcas | Published: 2024-04-01 | Updated: 2024-04-02
LLMセキュリティ
プライバシー保護
プライバシー保護手法

To Err is Machine: Vulnerability Detection Challenges LLM Reasoning

Authors: Benjamin Steenhoek, Md Mahbubur Rahman, Monoshi Kumar Roy, Mirza Sanjida Alam, Hengbo Tong, Swarna Das, Earl T. Barr, Wei Le | Published: 2024-03-25 | Updated: 2025-01-07
DoS対策
LLMセキュリティ
プロンプトインジェクション

Large Language Models for Blockchain Security: A Systematic Literature Review

Authors: Zheyuan He, Zihao Li, Sen Yang, He Ye, Ao Qiao, Xiaosong Zhang, Xiapu Luo, Ting Chen | Published: 2024-03-21 | Updated: 2025-03-24
LLMセキュリティ
アルゴリズム
ブロックチェーン技術

Large language models in 6G security: challenges and opportunities

Authors: Tri Nguyen, Huong Nguyen, Ahmad Ijaz, Saeid Sheikhi, Athanasios V. Vasilakos, Panos Kostakos | Published: 2024-03-18
LLMセキュリティ
サイバーセキュリティ
分散型LLMアーキテクチャ

What Was Your Prompt? A Remote Keylogging Attack on AI Assistants

Authors: Roy Weiss, Daniel Ayzenshteyn, Guy Amit, Yisroel Mirsky | Published: 2024-03-14
LLMセキュリティ
トークンの処理と収集
トークン収集手法

CodeAttack: Revealing Safety Generalization Challenges of Large Language Models via Code Completion

Authors: Qibing Ren, Chang Gao, Jing Shao, Junchi Yan, Xin Tan, Wai Lam, Lizhuang Ma | Published: 2024-03-12 | Updated: 2024-09-14
LLMセキュリティ
コード生成
プロンプトインジェクション

Fuzzing BusyBox: Leveraging LLM and Crash Reuse for Embedded Bug Unearthing

Authors: Asmita, Yaroslav Oliinyk, Michael Scott, Ryan Tsang, Chongzhou Fang, Houman Homayoun | Published: 2024-03-06
LLMセキュリティ
ファジング
初期シード生成

AutoAttacker: A Large Language Model Guided System to Implement Automatic Cyber-attacks

Authors: Jiacen Xu, Jack W. Stokes, Geoff McDonald, Xuesong Bai, David Marshall, Siyue Wang, Adith Swaminathan, Zhou Li | Published: 2024-03-02
LLMセキュリティ
プロンプトインジェクション
攻撃手法

Making Them Ask and Answer: Jailbreaking Large Language Models in Few Queries via Disguise and Reconstruction

Authors: Tong Liu, Yingjie Zhang, Zhe Zhao, Yinpeng Dong, Guozhu Meng, Kai Chen | Published: 2024-02-28 | Updated: 2024-06-10
LLMセキュリティ
LLM性能評価
プロンプトインジェクション