These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
Large language models (LLMs) are increasingly integrated into real-time
machine learning applications, where safeguarding user privacy is paramount.
Traditional differential privacy mechanisms often struggle to balance privacy
and accuracy, particularly in fast-changing environments with continuously
flowing data. To address these issues, we introduce Scalable Differential
Privacy (SDP), a framework tailored for real-time machine learning that
emphasizes both robust privacy guarantees and enhanced model performance. SDP
employs a hierarchical architecture to facilitate efficient noise aggregation
across various learning agents. By integrating adaptive noise scheduling and
gradient compression methods, our approach minimizes performance degradation
while ensuring significant privacy protection. Extensive experiments on diverse
datasets reveal that SDP maintains high accuracy levels while applying
differential privacy effectively, showcasing its suitability for deployment in
sensitive domains. This advancement points towards the potential for widespread
adoption of privacy-preserving techniques in machine learning workflows.