Open Access iconOpen Access

ARTICLE

crossmark

ANNDRA-IoT: A Deep Learning Approach for Optimal Resource Allocation in Internet of Things Environments

Abdullah M. Alqahtani1,*, Kamran Ahmad Awan2, Abdulaziz Almaleh3, Osama Aletri4

1 Department of Electrical and Electronic Engineering, College of Engineering and Computer Science, Jazan University, Jazan, 45142, Saudi Arabia
2 Department of Information Technology, The University of Haripur, Haripur, 22620, Pakistan
3 College of Computer Science, King Khalid University, Abha, 62529, Saudi Arabia
4 Department of Computing, College of Engineering and Computing, Umm Al-Qura University, Makkah, 21955, Saudi Arabia

* Corresponding Author: Abdullah M. Alqahtani. Email: email

(This article belongs to the Special Issue: Leveraging AI and ML for QoS Improvement in Intelligent Programmable Networks)

Computer Modeling in Engineering & Sciences 2025, 142(3), 3155-3179. https://doi.org/10.32604/cmes.2025.061472

Abstract

Efficient resource management within Internet of Things (IoT) environments remains a pressing challenge due to the increasing number of devices and their diverse functionalities. This study introduces a neural network-based model that uses Long-Short-Term Memory (LSTM) to optimize resource allocation under dynamically changing conditions. Designed to monitor the workload on individual IoT nodes, the model incorporates long-term data dependencies, enabling adaptive resource distribution in real time. The training process utilizes Min-Max normalization and grid search for hyperparameter tuning, ensuring high resource utilization and consistent performance. The simulation results demonstrate the effectiveness of the proposed method, outperforming the state-of-the-art approaches, including Dynamic and Efficient Enhanced Load-Balancing (DEELB), Optimized Scheduling and Collaborative Active Resource-management (OSCAR), Convolutional Neural Network with Monarch Butterfly Optimization (CNN-MBO), and Autonomic Workload Prediction and Resource Allocation for Fog (AWPR-FOG). For example, in scenarios with low system utilization, the model achieved a resource utilization efficiency of 95% while maintaining a latency of just 15 ms, significantly exceeding the performance of comparative methods.

Keywords

Internet of things; resource optimization; deep learning; optimal resource allocation; neural network; efficiency

Cite This Article

APA Style
Alqahtani, A.M., Awan, K.A., Almaleh, A., Aletri, O. (2025). Anndra-iot: A deep learning approach for optimal resource allocation in internet of things environments. Computer Modeling in Engineering & Sciences, 142(3), 3155–3179. https://doi.org/10.32604/cmes.2025.061472
Vancouver Style
Alqahtani AM, Awan KA, Almaleh A, Aletri O. Anndra-iot: A deep learning approach for optimal resource allocation in internet of things environments. Comput Model Eng Sci. 2025;142(3):3155–3179. https://doi.org/10.32604/cmes.2025.061472
IEEE Style
A. M. Alqahtani, K. A. Awan, A. Almaleh, and O. Aletri, “ANNDRA-IoT: A Deep Learning Approach for Optimal Resource Allocation in Internet of Things Environments,” Comput. Model. Eng. Sci., vol. 142, no. 3, pp. 3155–3179, 2025. https://doi.org/10.32604/cmes.2025.061472



cc Copyright © 2025 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 138

    View

  • 146

    Download

  • 0

    Like

Share Link