LLM Performance Evaluation

Focus Session: LLM4PQC — An Agentic Framework for Accurate and Efficient Synthesis of PQC Cores

Authors: Buddhi Perera, Zeng Wang, Weihua Xiao, Mohammed Nabeel, Ozgur Sinanoglu, Johann Knechtel, Ramesh Karri | Published: 2026-02-10
LLM Performance Evaluation
Hardware Accelerator
Prompt leaking

A Behavioral Fingerprint for Large Language Models: Provenance Tracking via Refusal Vectors

Authors: Zhenyu Xu, Victor S. Sheng | Published: 2026-02-10
Disabling Safety Mechanisms of LLM
LLM Performance Evaluation
evaluation metrics

LLMAC: A Global and Explainable Access Control Framework with Large Language Model

Authors: Sharif Noor Zisad, Ragib Hasan | Published: 2026-02-10
LLM Performance Evaluation
Poisoning attack on RAG
アクセス制御モデル

Towards Real-World Industrial-Scale Verification: LLM-Driven Theorem Proving on seL4

Authors: Jianyu Zhang, Fuyuan Zhang, Jiayi Lu, Jilin Hu, Xiaoyi Yin, Long Zhang, Feng Yang, Yongwang Zhao | Published: 2026-02-09
LLM Performance Evaluation
Program Understanding
Transparency and Verification

InfiCoEvalChain: A Blockchain-Based Decentralized Framework for Collaborative LLM Evaluation

Authors: Yifan Yang, Jinjia Li, Kunxi Li, Puhao Zheng, Yuanyi Wang, Zheyan Qu, Yang Yu, Jianmin Wu, Ming Li, Hongxia Yang | Published: 2026-02-09
LLM Performance Evaluation
Incentive Mechanism
Model evaluation methods

BadTemplate: A Training-Free Backdoor Attack via Chat Template Against Large Language Models

Authors: Zihan Wang, Hongwei Li, Rui Zhang, Wenbo Jiang, Guowen Xu | Published: 2026-02-05
LLM Performance Evaluation
データ毒性
Large Language Model

SynAT: Enhancing Security Knowledge Bases via Automatic Synthesizing Attack Tree from Crowd Discussions

Authors: Ziyou Jiang, Lin Shi, Guowei Yang, Xuyan Ma, Fenglong Li, Qing Wang | Published: 2026-02-05
LLM Performance Evaluation
Safety of Data Generation
攻撃ツリー合成

Hallucination-Resistant Security Planning with a Large Language Model

Authors: Kim Hammar, Tansu Alpcan, Emil Lupu | Published: 2026-02-05
LLM Performance Evaluation
Hallucination
Detection of Hallucinations

How Few-shot Demonstrations Affect Prompt-based Defenses Against LLM Jailbreak Attacks

Authors: Yanshu Wang, Shuaishuai Yang, Jingjing He, Tong Yang | Published: 2026-02-04
LLM Performance Evaluation
Prompt Injection
Large Language Model

LogicScan: An LLM-driven Framework for Detecting Business Logic Vulnerabilities in Smart Contracts

Authors: Jiaqi Gao, Zijian Zhang, Yuqiang Sun, Ye Liu, Chengwei Liu, Han Liu, Yi Li, Yang Liu | Published: 2026-02-03
LLM Performance Evaluation
スマートコントラクト攻撃
Prompt leaking