LLM Performance Evaluation

Superficial Safety Alignment Hypothesis

Authors: Jianwei Li, Jung-Eun Kim | Published: 2024-10-07
LLM Performance Evaluation
Safety Alignment

DiDOTS: Knowledge Distillation from Large-Language-Models for Dementia Obfuscation in Transcribed Speech

Authors: Dominika Woszczyk, Soteris Demetriou | Published: 2024-10-05
LLM Performance Evaluation
Privacy Protection

A Watermark for Black-Box Language Models

Authors: Dara Bahri, John Wieting, Dana Alon, Donald Metzler | Published: 2024-10-02
LLM Performance Evaluation
Watermarking
Watermark Evaluation

PathSeeker: Exploring LLM Security Vulnerabilities with a Reinforcement Learning-Based Jailbreak Approach

Authors: Zhihao Lin, Wei Ma, Mingyi Zhou, Yanjie Zhao, Haoyu Wang, Yang Liu, Jun Wang, Li Li | Published: 2024-09-21 | Updated: 2024-10-03
LLM Performance Evaluation
Prompt Injection

CLNX: Bridging Code and Natural Language for C/C++ Vulnerability-Contributing Commits Identification

Authors: Zeqing Qin, Yiwei Wu, Lansheng Han | Published: 2024-09-11
LLM Performance Evaluation
Program Analysis
Prompt Injection

DrLLM: Prompt-Enhanced Distributed Denial-of-Service Resistance Method with Large Language Models

Authors: Zhenyu Yin, Shang Liu, Guangyuan Xu | Published: 2024-09-11 | Updated: 2025-01-13
DDoS Attack Detection
LLM Performance Evaluation
Prompt Injection

LLM-Enhanced Software Patch Localization

Authors: Jinhong Yu, Yi Chen, Di Tang, Xiaozhong Liu, XiaoFeng Wang, Chen Wu, Haixu Tang | Published: 2024-09-10 | Updated: 2024-09-13
LLM Performance Evaluation
Understanding Commit Content
Evaluation Method

Using Large Language Models for Template Detection from Security Event Logs

Authors: Risto Vaarandi, Hayretdin Bahsi | Published: 2024-09-08 | Updated: 2025-04-14
LLM Performance Evaluation
Algorithm
Evaluation Method

RACONTEUR: A Knowledgeable, Insightful, and Portable LLM-Powered Shell Command Explainer

Authors: Jiangyi Deng, Xinfeng Li, Yanjiao Chen, Yijie Bai, Haiqin Weng, Yan Liu, Tao Wei, Wenyuan Xu | Published: 2024-09-03
LLM Performance Evaluation
Cybersecurity
Prompt Injection

The Dark Side of Human Feedback: Poisoning Large Language Models via User Inputs

Authors: Bocheng Chen, Hanqing Guo, Guangjing Wang, Yuanda Wang, Qiben Yan | Published: 2024-09-01
LLM Performance Evaluation
Prompt Injection
Poisoning