Files / Emerging technologies

Artificial Intelligence Cybersecurity

A systematic collection of studies on malware detection, adversarial attacks, and reinforcement learning applications, covering in-depth technical analysis from theoretical frameworks to real-world security scenarios.

Detail

Published

22/12/2025

Key Chapter Title List

  1. Adversarial Malware and Benign Sample Generation Based on Reinforcement Learning
  2. Application of Auxiliary Classifier Generative Adversarial Networks in Malware Analysis
  3. Evaluating the Robustness of Image-Based Malware Classifiers Through Minor Perturbation Techniques
  4. Detecting Botnets Using Deep Learning and Network Flow Analysis
  5. Interpretability Analysis of Machine Learning-Based Malware Detection Results Using Rule Sets
  6. Mobile Malware Detection Based on Consortium Blockchain
  7. Application of the BERT Model in Malware Classification
  8. Machine Learning Methods for Malware Evolution Detection

File Introduction

This research collection focuses on the cutting-edge applications and challenges of artificial intelligence technology in cybersecurity, particularly in the field of malware detection and adversarial scenarios. As machine learning increasingly becomes a core component of cybersecurity defense systems, the inherent vulnerabilities of its models, the black-box nature of their decision-making processes, and their robustness issues in adversarial environments have become key bottlenecks limiting their effectiveness and trustworthiness. Through multiple in-depth technical papers, this report systematically explores how to utilize and defend AI technology to build more secure, reliable, and interpretable malware detection systems.

The core issues revolve around two main themes: first, the attack perspective, which involves using advanced AI technologies (such as reinforcement learning and generative adversarial networks) to generate adversarial malware samples capable of evading existing detection mechanisms; second, the defense and interpretation perspective, which focuses on enhancing the resistance of detection models to such attacks and improving the transparency and comprehensibility of their decision-making processes. Specific research covers various malware carriers, from binary executable files (PE format) to Android applications (APK/DEX), and employs multiple analytical methods, including static feature analysis, image-based representation, and network traffic analysis.

In terms of attack methodology, the report elaborates on a black-box attack framework based on reinforcement learning. This framework successfully deceives mainstream machine learning classifiers such as LightGBM and MalConv by making functionality-preserving modifications to executable files, achieving a notable evasion rate against some real antivirus software. Simultaneously, the research explores the use of Generative Adversarial Networks (GANs) to generate realistic fake malware images to test the discriminative capabilities and potential weaknesses of classifiers. On the other hand, for image-based classifiers on the Android platform, the report demonstrates that simple code-level perturbations (such as inserting no-operation instructions or redundant strings) can significantly reduce model classification accuracy, revealing the fragility of such methods.

In the areas of defense, detection, and interpretability, the report proposes a botnet detection method based on network flow pattern analysis and deep learning. By extracting statistical features such as the periodicity of communications, it achieves effective identification of command-and-control servers. To demystify the black box of machine learning models, the research extensively applies rule-based classification methods (such as the I-REP and RIPPER algorithms), attempting to transform complex model predictions into human-understandable decision rule lists. It also proposes quantitative evaluation metrics like interpretability entropy to measure the difficulty of explaining the output results of different machine learning models.

This collection integrates multiple independent yet complementary in-depth studies, providing cybersecurity researchers, AI practitioners, and anti-malware engineers with a comprehensive technical reference on the latest advancements, methods, challenges, and solutions in the intersecting field of AI and security. Its content includes not only specific algorithm implementations, experimental designs, and result analyses but also offers profound insights into the reliability, interpretability, and future development paths of AI models in real adversarial environments.