The rapid growth of blockchain technology has driven the widespread adoption
of smart contracts. However, their inherent vulnerabilities have led to
significant financial losses. Traditional auditing methods, while essential,
struggle to keep pace with the increasing complexity and scale of smart
contracts. Large Language Models (LLMs) offer promising capabilities for
automating vulnerability detection, but their adoption is often limited by high
computational costs. Although prior work has explored leveraging large models
through agents or workflows, relatively little attention has been given to
improving the performance of smaller, fine-tuned models--a critical factor for
achieving both efficiency and data privacy. In this paper, we introduce
HKT-SmartAudit, a framework for developing lightweight models optimized for
smart contract auditing. It features a multi-stage knowledge distillation
pipeline that integrates classical distillation, external domain knowledge, and
reward-guided learning to transfer high-quality insights from large teacher
models. A single-task learning strategy is employed to train compact student
models that maintain high accuracy and robustness while significantly reducing
computational overhead. Experimental results show that our distilled models
outperform both commercial tools and larger models in detecting complex
vulnerabilities and logical flaws, offering a practical, secure, and scalable
solution for smart contract auditing. The source code is available at Github
repository.