These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
Adversarial attacks are a potential threat to machine learning models by
causing incorrect predictions through imperceptible perturbations to the input
data. While these attacks have been extensively studied in unstructured data
like images, applying them to tabular data, poses new challenges. These
challenges arise from the inherent heterogeneity and complex feature
interdependencies in tabular data, which differ from the image data. To account
for this distinction, it is necessary to establish tailored imperceptibility
criteria specific to tabular data. However, there is currently a lack of
standardised metrics for assessing the imperceptibility of adversarial attacks
on tabular data. To address this gap, we propose a set of key properties and
corresponding metrics designed to comprehensively characterise imperceptible
adversarial attacks on tabular data. These are: proximity to the original
input, sparsity of altered features, deviation from the original data
distribution, sensitivity in perturbing features with narrow distribution,
immutability of certain features that should remain unchanged, feasibility of
specific feature values that should not go beyond valid practical ranges, and
feature interdependencies capturing complex relationships between data
attributes. We evaluate the imperceptibility of five adversarial attacks,
including both bounded attacks and unbounded attacks, on tabular data using the
proposed imperceptibility metrics. The results reveal a trade-off between the
imperceptibility and effectiveness of these attacks. The study also identifies
limitations in current attack algorithms, offering insights that can guide
future research in the area. The findings gained from this empirical analysis
provide valuable direction for enhancing the design of adversarial attack
algorithms, thereby advancing adversarial machine learning on tabular data.