AIセキュリティポータル K Program
Explainability-Informed Targeted Malware Misclassification
Share
Abstract
In recent years, there has been a surge in malware attacks across critical infrastructures, requiring further research and development of appropriate response and remediation strategies in malware detection and classification. Several works have used machine learning models for malware classification into categories, and deep neural networks have shown promising results. However, these models have shown its vulnerabilities against intentionally crafted adversarial attacks, which yields misclassification of a malicious file. Our paper explores such adversarial vulnerabilities of neural network based malware classification system in the dynamic and online analysis environments. To evaluate our approach, we trained Feed Forward Neural Networks (FFNN) to classify malware categories based on features obtained from dynamic and online analysis environments. We use the state-of-the-art method, SHapley Additive exPlanations (SHAP), for the feature attribution for malware classification, to inform the adversarial attackers about the features with significant importance on classification decision. Using the explainability-informed features, we perform targeted misclassification adversarial white-box evasion attacks using the Fast Gradient Sign Method (FGSM) and Projected Gradient Descent (PGD) attacks against the trained classifier. Our results demonstrated high evasion rate for some instances of attacks, showing a clear vulnerability of a malware classifier for such attacks. We offer recommendations for a balanced approach and a benchmark for much-needed future research into evasion attacks against malware classifiers, and develop more robust and trustworthy solutions.
Malware detection with deep neural network using process behavior
S. Tobiyama, Y. Yamaguchi, H. Shimada, T. Ikuse, T. Yagi
Published: 2016
Didroid: Android malware classification and characterization using deep image learning
A. Rahali, et al.
Published: 2020
Entroplyzer: Android malware classification and characterization using entropy analysis of dynamic characteristics
D. S. Keyes, B. Li, G. Kaur, A. H. Lashkari, et al.
Published: 2021
Analyzing Machine Learning Approaches for Online Malware Detection in Cloud
Jeffrey C Kimmell, Mahmoud Abdelsalam, Maanak Gupta
Published: 5.20.2021
COPYCAT: Practical Adversarial Attacks on Visualization-Based Malware Detection
Aminollah Khormali, Ahmed Abusnaina, Songqing Chen, DaeHun Nyang, Aziz Mohaisen
Published: 9.21.2019
Optimization of code caves in malware binaries to evade machine learning detectors
J. Yuste, E. G. Pardo, J. Tapiador
Published: 2022
Exploiting windows pe structure for adversarial malware evasion attacks
K. Aryal, M. Gupta, M. Abdelsalam
Published: 2023
Intra-section code cave injection for adversarial evasion attacks on windows pe malware file
K. Aryal, M. Gupta, M. Abdelsalam, M. Saleh
Published: 2024
Attack and defense of dynamic analysis-based, adversarial neural malware detection models
J. W. Stokes, et al.
Published: 2018
Deceiving portable executable malware classifiers into targeted misclassification with practical adversarial examples
Y. Kucuk, G. Yan
Published: 2020
Mitigating adversarial evasion attacks of ransomware using ensemble learning
U. Ahmed, J. C.-W. Lin, G. Srivastava
Published: 2022
Mitigating malicious adversaries evasion attacks in industrial internet of things
H. Rafiq, et al.
Published: 2023
A Unified Approach to Interpreting Model Predictions
Scott Lundberg, Su-In Lee
Published: 5.23.2017
Radar: A real-world dataset for ai powered run-time detection of cyber-attacks
S. Karapoola, N. Singh, C. Rebeiro, K. V.
Published: 2022
Explaining and harnessing adversarial examples
Goodfellow, I. J., Shlens, J., Szegedy, C.
Published: 2015
Towards Deep Learning Models Resistant to Adversarial Attacks
Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, Adrian Vladu
Published: 6.20.2017
Share