TY - GEN
T1 - Bio-inspired Hyperparameter Optimization for Deep Learning Malware Detection
T2 - 2025 The 8th International Conference on Computational Intelligence and Intelligent Systems, CIIS 2025
AU - Mokone, Kgosietsile D.
AU - Sithungu, Siphesihle P.
AU - Ehlers, Elizabeth M.
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s).
PY - 2026/2/2
Y1 - 2026/2/2
N2 - The performance of deep learning malware classifiers depends heavily on hyperparameter configuration. Deep learning malware detection benefits from principled hyperparameter optimization. This study evaluates six bio inspired optimizers across four neural architectures and three widely used malware data representations. Artificial Immune System methods include Clonal Selection, Immune Network, and a Dendritic Cell inspired optimizer. Evolutionary methods include Genetic Algorithm, Particle Swarm, and Differential Evolution. Baselines for each deep learning model were implemented using domain informed choices rather than naïve defaults to avoid overstating optimization gains and to reflect realistic practitioner practice. All runs were under a strictly matched evaluation budget (g1/460 function evaluations per run) with five independent seeds per permutation. On API call sequences, LSTM produced the highest test accuracy near 0.991, with Differential Evolution and Particle Swarm frequently leading, while Transformer on the same dataset reached about 0.990. On PE headers, ANN achieved about 0.989 with Differential Evolution and Immune Network. Opcode n grams formed the most challenging setting, with the best CNN accuracy near 0.75 under multi class family labels. Artificial Immune System algorithms often reached the highest accuracies but typically at greater runtime cost, whereas Evolutionary Algorithms, especially Differential Evolution and Particle Swarm, were faster to strong solutions. under a strictly matched evaluation budget (g1/460 function evaluations per run) with five independent seeds per permutation. The optimizers all perform comparatively well to each other on each dataset, with no statistical difference in test accuracy. The experiment also quantify accuracy versus computing tradeoffs, highlight dataset model synergies, and provide guidance on selecting model optimizer pairs for deployment-oriented scenarios. The research notes ceiling effects on API/PE and a modest, matched budget; future work will examine larger budgets and more challenging datasets to further probe optimizer differences under controlled evaluation.
AB - The performance of deep learning malware classifiers depends heavily on hyperparameter configuration. Deep learning malware detection benefits from principled hyperparameter optimization. This study evaluates six bio inspired optimizers across four neural architectures and three widely used malware data representations. Artificial Immune System methods include Clonal Selection, Immune Network, and a Dendritic Cell inspired optimizer. Evolutionary methods include Genetic Algorithm, Particle Swarm, and Differential Evolution. Baselines for each deep learning model were implemented using domain informed choices rather than naïve defaults to avoid overstating optimization gains and to reflect realistic practitioner practice. All runs were under a strictly matched evaluation budget (g1/460 function evaluations per run) with five independent seeds per permutation. On API call sequences, LSTM produced the highest test accuracy near 0.991, with Differential Evolution and Particle Swarm frequently leading, while Transformer on the same dataset reached about 0.990. On PE headers, ANN achieved about 0.989 with Differential Evolution and Immune Network. Opcode n grams formed the most challenging setting, with the best CNN accuracy near 0.75 under multi class family labels. Artificial Immune System algorithms often reached the highest accuracies but typically at greater runtime cost, whereas Evolutionary Algorithms, especially Differential Evolution and Particle Swarm, were faster to strong solutions. under a strictly matched evaluation budget (g1/460 function evaluations per run) with five independent seeds per permutation. The optimizers all perform comparatively well to each other on each dataset, with no statistical difference in test accuracy. The experiment also quantify accuracy versus computing tradeoffs, highlight dataset model synergies, and provide guidance on selecting model optimizer pairs for deployment-oriented scenarios. The research notes ceiling effects on API/PE and a modest, matched budget; future work will examine larger budgets and more challenging datasets to further probe optimizer differences under controlled evaluation.
KW - artificial immune systems
KW - deep learning
KW - evolutionary algorithms
KW - hyperparameter optimization
KW - Malware detection
UR - https://www.scopus.com/pages/publications/105030546199
U2 - 10.1145/3787256.3787278
DO - 10.1145/3787256.3787278
M3 - Conference contribution
AN - SCOPUS:105030546199
T3 - CIIS 2025 - 2025 the 8th International Conference on Computational Intelligence and Intelligent Systems
SP - 145
EP - 154
BT - CIIS 2025 - 2025 the 8th International Conference on Computational Intelligence and Intelligent Systems
PB - Association for Computing Machinery, Inc
Y2 - 21 November 2025 through 23 November 2025
ER -