These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
Machine learning has revolutionized numerous domains, playing a crucial role
in driving advancements and enabling data-centric processes. The significance
of data in training models and shaping their performance cannot be overstated.
Recent research has highlighted the heterogeneous impact of individual data
samples, particularly the presence of valuable data that significantly
contributes to the utility and effectiveness of machine learning models.
However, a critical question remains unanswered: are these valuable data
samples more vulnerable to machine learning attacks? In this work, we
investigate the relationship between data importance and machine learning
attacks by analyzing five distinct attack types. Our findings reveal notable
insights. For example, we observe that high importance data samples exhibit
increased vulnerability in certain attacks, such as membership inference and
model stealing. By analyzing the linkage between membership inference
vulnerability and data importance, we demonstrate that sample characteristics
can be integrated into membership metrics by introducing sample-specific
criteria, therefore enhancing the membership inference performance. These
findings emphasize the urgent need for innovative defense mechanisms that
strike a balance between maximizing utility and safeguarding valuable data
against potential exploitation.