Adversarial Attacks on Featureless Deep Learning Malicious URLs Detection

Detecting malicious Uniform Resource Locators (URLs) is crucially important to prevent attackers from committing cybercrimes. Recent researches have investigated the role of machine learning (ML) models to detect malicious URLs. By using ML algorithms, first, the features of URLs are extracted, and...

Full description

Saved in:
Bibliographic Details
Published inComputers, materials & continua Vol. 68; no. 1; pp. 921 - 939
Main Authors Rasheed, Bader, Khan, Adil, M. Ahsan Kazmi, S., Hussain, Rasheed, Jalil Piran, Md, Young Suh, Doug
Format Journal Article
LanguageEnglish
Published Henderson Tech Science Press 01.01.2021
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Detecting malicious Uniform Resource Locators (URLs) is crucially important to prevent attackers from committing cybercrimes. Recent researches have investigated the role of machine learning (ML) models to detect malicious URLs. By using ML algorithms, first, the features of URLs are extracted, and then different ML models are trained. The limitation of this approach is that it requires manual feature engineering and it does not consider the sequential patterns in the URL. Therefore, deep learning (DL) models are used to solve these issues since they are able to perform featureless detection. Furthermore, DL models give better accuracy and generalization to newly designed URLs; however, the results of our study show that these models, such as any other DL models, can be susceptible to adversarial attacks. In this paper, we examine the robustness of these models and demonstrate the importance of considering this susceptibility before applying such detection systems in real-world solutions. We propose and demonstrate a black-box attack based on scoring functions with greedy search for the minimum number of perturbations leading to a misclassification. The attack is examined against different types of convolutional neural networks (CNN)-based URL classifiers and it causes a tangible decrease in the accuracy with more than 56% reduction in the accuracy of the best classifier (among the selected classifiers for this work). Moreover, adversarial training shows promising results in reducing the influence of the attack on the robustness of the model to less than 7% on average.
ISSN:1546-2226
1546-2218
1546-2226
DOI:10.32604/cmc.2021.015452