These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
Machine unlearning, a process enabling pre-trained models to remove the
influence of specific training samples, has attracted significant attention in
recent years. While extensive research has focused on developing efficient
unlearning strategies, the critical aspect of unlearning verification has been
largely overlooked. Existing verification methods mainly rely on machine
learning attack techniques, such as membership inference attacks (MIAs) or
backdoor attacks. However, these methods, not being formally designed for
verification purposes, exhibit limitations in robustness and only support a
small, predefined subset of samples. Moreover, dependence on prepared
sample-level modifications of MIAs or backdoor attacks restricts their
applicability in Machine Learning as a Service (MLaaS) environments. To address
these limitations, we propose a novel robustness verification scheme without
any prior modifications, and can support verification on a much larger set. Our
scheme employs an optimization-based method to recover the actual training
samples from the model. By comparative analysis of recovered samples extracted
pre- and post-unlearning, MLaaS users can verify the unlearning process. This
verification scheme, operating exclusively through model parameters, avoids the
need for any sample-level modifications prior to model training while
supporting verification on a much larger set and maintaining robustness. The
effectiveness of our proposed approach is demonstrated through theoretical
analysis and experiments involving diverse models on various datasets in
different scenarios.