INTERNATIONAL JOURNAL OF LATEST TECHNOLOGY IN ENGINEERING,
MANAGEMENT & APPLIED SCIENCE (IJLTEMAS)
ISSN 2278-2540 | DOI: 10.51583/IJLTEMAS | Volume XIV, Issue X, October 2025
Gradient Boosting Machines (GBM): GBM is an ensemble learning technique that builds a strong classifier by combining
multiple decision trees. It is highly effective for structured data, making it suitable for static features like API call frequencies,
entropy, and file headers. The training process begins with the preprocessed dataset obtained from the feature selection step. The
dataset is split into training, validation, and test subsets to prevent overfitting and ensure generalizability.
The GBM model is configured with parameters such as the number of estimators, learning rate, and tree depth. Hyperparameter
tuning is performed using grid or random search methods. The training process minimises the error iteratively, allowing the model
to assign higher weights to misclassified samples in each round. The model is then evaluated on the test set using metrics like
accuracy, precision, recall, F1-score, and AUC. These metrics provide insights into the model's classification performance, ensuring
it effectively distinguishes between malware families.
Convolutional Neural Networks (CNN): CNNs are deep learning models designed to process data with spatial hierarchies, such
as images or sequential data. For malware classification, features like opcode sequences, byte plots, or network activity can be
transformed into structured representations, such as images or matrices. This transformation allows CNNs to analyse complex
patterns and relationships inherent in malware behaviour.
The CNN architecture typically consists of convolutional layers for feature extraction, pooling layers for dimensionality reduction,
and fully connected layers for final classification. Dropout layers are used to prevent overfitting, and the model is trained using a
loss function, such as categorical cross-entropy, optimised with the Adam optimiser. The training process involves feeding the
transformed data into the model in batches, monitoring performance on the validation set to adjust learning rates, and stopping early
if the validation loss stops improving. Finally, the model is evaluated on the test data, providing accuracy, precision, and recall as
indicators of its effectiveness in malware classification.
Gated Recurrent Units (GRU): Gated Recurrent Units (GRUs) are an advanced variant of recurrent neural networks (RNNs)
designed to overcome the vanishing gradient problem commonly encountered in sequence modelling. Unlike traditional RNNs,
GRUs utilise a gating mechanism, comprising reset and update gates, that allows the network to retain or discard information over
long sequences efficiently. This architecture enables GRUs to capture long-term dependencies in sequential data while maintaining
lower computational complexity compared to Long Short-Term Memory (LSTM) networks. In the context of malware de-
obfuscation, GRUs are particularly advantageous because malware often exhibits sequential dependencies in execution traces, API
call logs, and opcode sequences. Obfuscation techniques attempt to break these recognisable patterns; however, GRUs can learn
contextual relationships across long behavioural sequences, making them resilient to such transformations.
Within the proposed hybrid system, the GRU is employed to analyse dynamic features extracted during reverse engineering, such
as runtime API call sequences and system interactions. By leveraging its gating mechanism, the GRU selectively focuses on the
most informative behaviours while ignoring redundant or noisy data. This sequential modelling complements the Gradient Boosting
Machine (GBM), which focuses on structured tabular attributes, and the Convolutional Neural Network (CNN), which captures
spatial features in binary or opcode image representations. The GRU is trained using binary cross-entropy loss for malware-versus-
benign classification, optimised with adaptive learning methods such as Adam. Its relatively lightweight architecture reduces
computational overhead, making it better suited than LSTMs for real-time or large-scale malware detection scenarios.
III. Results and Discussion
The proposed system integrated reverse engineering techniques with ensemble learning models comprising Gradient Boosting
Machine (GBM), Convolutional Neural Networks (CNN), and Gated Recurrent Units (GRU). Reverse engineering facilitated both
static and dynamic feature extraction from obfuscated malware samples, enabling the system to capture opcode patterns, API call
traces, and execution behaviors. These features were subsequently processed by the ensemble model, where each component
contributed distinct strengths. GBM proved effective in analysing structured tabular features such as opcode frequency distributions
and executable metadata. CNN successfully extracted spatial dependencies from binary visualisation of malware samples and
opcode images, while GRU captured sequential dependencies in API call traces and execution logs, making the system more
resilient to obfuscation and polymorphic malware. The integration of these models was achieved through a weighted voting
ensemble mechanism known as the stacking ensemble learning technique, ensuring that predictions incorporated the
complementary advantages of each model while minimising their individual weaknesses. When tested on a balanced dataset of
obfuscated malware and benign files, the ensemble system demonstrated superior performance compared to the individual
classifiers. Specifically, GBM achieved an accuracy of 92.6%, CNN reached 93.8%, and GRU achieved 95.1%. The ensemble,
however, surpassed these results by attaining 97.4% accuracy, 95.8% precision, and an F1-score of 96.1%. Furthermore, the
ensemble recorded an AUC-ROC score of 0.985 and reduced the false positive rate to 2.6%, representing a significant improvement
over the standalone models.
Mathematical Representation of the Performance Metrics
The standard terms used in evaluating the performance of a classification model are:
a. True Positives (TP): It is an outcome of classification where malware is correctly classified as malware.
b. True Negatives (TN): It is an outcome of classification where a benign file is correctly classified as benign.
Page 1310