These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
Large language models (LLMs), especially those based on the Transformer
architecture, have had a profound impact on various aspects of daily life, such
as natural language processing, content generation, research methodologies, and
more. Nevertheless, a crucial concern regarding the inference results of large
language models is the issue of security and privacy. Given that large language
models can generate results that may leak sensitive confidential or copyright
information in many scenarios, it is crucial to compute the attention matrix
with provable privacy guarantees, as attention is all you need.
In this work, we propose a novel and efficient algorithm for approximating
the attention matrix while providing differential privacy (DP) guarantees. To
achieve this, we build on recent advancements in fast attention computation and
differentially private matrix publishing.