These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
Decentralized applications (DApps) face significant security risks due to
vulnerabilities in smart contracts, with traditional detection methods
struggling to address emerging and machine-unauditable flaws. This paper
proposes a novel approach leveraging fine-tuned Large Language Models (LLMs) to
enhance smart contract vulnerability detection. We introduce a comprehensive
dataset of 215 real-world DApp projects (4,998 contracts), including
hard-to-detect logical errors like token price manipulation, addressing the
limitations of existing simplified benchmarks. By fine-tuning LLMs (Llama3-8B
and Qwen2-7B) with Full-Parameter Fine-Tuning (FFT) and Low-Rank Adaptation
(LoRA), our method achieves superior performance, attaining an F1-score of 0.83
with FFT and data augmentation via Random Over Sampling (ROS). Comparative
experiments demonstrate significant improvements over prompt-based LLMs and
state-of-the-art tools. Notably, the approach excels in detecting
non-machine-auditable vulnerabilities, achieving 0.97 precision and 0.68 recall
for price manipulation flaws. The results underscore the effectiveness of
domain-specific LLM fine-tuning and data augmentation in addressing real-world
DApp security challenges, offering a robust solution for blockchain ecosystem
protection.
External Datasets
215 DApp projects (4,998 contracts)
23 contracts with token price manipulation vulnerabilities