Home / Advanced Search

  • Title/Keywords

  • Author/Affliations

  • Journal

  • Article Type

  • Start Year

  • End Year

Update SearchingClear
  • Articles
  • Online
Search Results (179)
  • Open Access

    ARTICLE

    A Hybrid Approach to Software Testing Efficiency: Stacked Ensembles and Deep Q-Learning for Test Case Prioritization and Ranking

    Anis Zarrad1, Thomas Armstrong2, Jaber Jemai3,*

    CMC-Computers, Materials & Continua, Vol.86, No.3, 2026, DOI:10.32604/cmc.2025.072768 - 12 January 2026

    Abstract Test case prioritization and ranking play a crucial role in software testing by improving fault detection efficiency and ensuring software reliability. While prioritization selects the most relevant test cases for optimal coverage, ranking further refines their execution order to detect critical faults earlier. This study investigates machine learning techniques to enhance both prioritization and ranking, contributing to more effective and efficient testing processes. We first employ advanced feature engineering alongside ensemble models, including Gradient Boosted, Support Vector Machines, Random Forests, and Naive Bayes classifiers to optimize test case prioritization, achieving an accuracy score of 0.98847More >

  • Open Access

    ARTICLE

    Research on UAV–MEC Cooperative Scheduling Algorithms Based on Multi-Agent Deep Reinforcement Learning

    Yonghua Huo1,2, Ying Liu1,*, Anni Jiang3, Yang Yang3

    CMC-Computers, Materials & Continua, Vol.86, No.3, 2026, DOI:10.32604/cmc.2025.072681 - 12 January 2026

    Abstract With the advent of sixth-generation mobile communications (6G), space–air–ground integrated networks have become mainstream. This paper focuses on collaborative scheduling for mobile edge computing (MEC) under a three-tier heterogeneous architecture composed of mobile devices, unmanned aerial vehicles (UAVs), and macro base stations (BSs). This scenario typically faces fast channel fading, dynamic computational loads, and energy constraints, whereas classical queuing-theoretic or convex-optimization approaches struggle to yield robust solutions in highly dynamic settings. To address this issue, we formulate a multi-agent Markov decision process (MDP) for an air–ground-fused MEC system, unify link selection, bandwidth/power allocation, and task… More >

  • Open Access

    ARTICLE

    DRL-Based Task Scheduling and Trajectory Control for UAV-Assisted MEC Systems

    Sai Xu1,*, Jun Liu1,*, Shengyu Huang1, Zhi Li2

    CMC-Computers, Materials & Continua, Vol.86, No.3, 2026, DOI:10.32604/cmc.2025.071865 - 12 January 2026

    Abstract In scenarios where ground-based cloud computing infrastructure is unavailable, unmanned aerial vehicles (UAVs) act as mobile edge computing (MEC) servers to provide on-demand computation services for ground terminals. To address the challenge of jointly optimizing task scheduling and UAV trajectory under limited resources and high mobility of UAVs, this paper presents PER-MATD3, a multi-agent deep reinforcement learning algorithm with prioritized experience replay (PER) into the Centralized Training with Decentralized Execution (CTDE) framework. Specifically, PER-MATD3 enables each agent to learn a decentralized policy using only local observations during execution, while leveraging a shared replay buffer with More >

  • Open Access

    ARTICLE

    AquaTree: Deep Reinforcement Learning-Driven Monte Carlo Tree Search for Underwater Image Enhancement

    Chao Li1,3,#, Jianing Wang1,3,#, Caichang Ding2,*, Zhiwei Ye1,3

    CMC-Computers, Materials & Continua, Vol.86, No.3, 2026, DOI:10.32604/cmc.2025.071242 - 12 January 2026

    Abstract Underwater images frequently suffer from chromatic distortion, blurred details, and low contrast, posing significant challenges for enhancement. This paper introduces AquaTree, a novel underwater image enhancement (UIE) method that reformulates the task as a Markov Decision Process (MDP) through the integration of Monte Carlo Tree Search (MCTS) and deep reinforcement learning (DRL). The framework employs an action space of 25 enhancement operators, strategically grouped for basic attribute adjustment, color component balance, correction, and deblurring. Exploration within MCTS is guided by a dual-branch convolutional network, enabling intelligent sequential operator selection. Our core contributions include: (1) a More >

  • Open Access

    ARTICLE

    Beyond Wi-Fi 7: Enhanced Decentralized Wireless Local Area Networks with Federated Reinforcement Learning

    Rashid Ali1,*, Alaa Omran Almagrabi2,3

    CMC-Computers, Materials & Continua, Vol.86, No.3, 2026, DOI:10.32604/cmc.2025.070224 - 12 January 2026

    Abstract Wi-Fi technology has evolved significantly since its introduction in 1997, advancing to Wi-Fi 6 as the latest standard, with Wi-Fi 7 currently under development. Despite these advancements, integrating machine learning into Wi-Fi networks remains challenging, especially in decentralized environments with multiple access points (mAPs). This paper is a short review that summarizes the potential applications of federated reinforcement learning (FRL) across eight key areas of Wi-Fi functionality, including channel access, link adaptation, beamforming, multi-user transmissions, channel bonding, multi-link operation, spatial reuse, and multi-basic servic set (multi-BSS) coordination. FRL is highlighted as a promising framework for More >

  • Open Access

    ARTICLE

    A Deep Reinforcement Learning-Based Partitioning Method for Power System Parallel Restoration

    Changcheng Li1,2, Weimeng Chang1,2, Dahai Zhang1,*, Jinghan He1

    Energy Engineering, Vol.123, No.1, 2026, DOI:10.32604/ee.2025.069389 - 27 December 2025

    Abstract Effective partitioning is crucial for enabling parallel restoration of power systems after blackouts. This paper proposes a novel partitioning method based on deep reinforcement learning. First, the partitioning decision process is formulated as a Markov decision process (MDP) model to maximize the modularity. Corresponding key partitioning constraints on parallel restoration are considered. Second, based on the partitioning objective and constraints, the reward function of the partitioning MDP model is set by adopting a relative deviation normalization scheme to reduce mutual interference between the reward and penalty in the reward function. The soft bonus scaling mechanism… More >

  • Open Access

    REVIEW

    Implementation of Human-AI Interaction in Reinforcement Learning: Literature Review and Case Studies

    Shaoping Xiao1,*, Zhaoan Wang1, Junchao Li2, Caden Noeller1, Jiefeng Jiang3, Jun Wang4

    CMC-Computers, Materials & Continua, Vol.86, No.2, pp. 1-62, 2026, DOI:10.32604/cmc.2025.072146 - 09 December 2025

    Abstract The integration of human factors into artificial intelligence (AI) systems has emerged as a critical research frontier, particularly in reinforcement learning (RL), where human-AI interaction (HAII) presents both opportunities and challenges. As RL continues to demonstrate remarkable success in model-free and partially observable environments, its real-world deployment increasingly requires effective collaboration with human operators and stakeholders. This article systematically examines HAII techniques in RL through both theoretical analysis and practical case studies. We establish a conceptual framework built upon three fundamental pillars of effective human-AI collaboration: computational trust modeling, system usability, and decision understandability. Our… More >

  • Open Access

    ARTICLE

    A Multi-Objective Adaptive Car-Following Framework for Autonomous Connected Vehicles with Deep Reinforcement Learning

    Abu Tayab1,*, Yanwen Li1, Ahmad Syed2, Ghanshyam G. Tejani3,4,*, Doaa Sami Khafaga5, El-Sayed M. El-kenawy6, Amel Ali Alhussan7, Marwa M. Eid8,9

    CMC-Computers, Materials & Continua, Vol.86, No.2, pp. 1-27, 2026, DOI:10.32604/cmc.2025.070583 - 09 December 2025

    Abstract Autonomous connected vehicles (ACV) involve advanced control strategies to effectively balance safety, efficiency, energy consumption, and passenger comfort. This research introduces a deep reinforcement learning (DRL)-based car-following (CF) framework employing the Deep Deterministic Policy Gradient (DDPG) algorithm, which integrates a multi-objective reward function that balances the four goals while maintaining safe policy learning. Utilizing real-world driving data from the highD dataset, the proposed model learns adaptive speed control policies suitable for dynamic traffic scenarios. The performance of the DRL-based model is evaluated against a traditional model predictive control-adaptive cruise control (MPC-ACC) controller. Results show that the… More >

  • Open Access

    ARTICLE

    Artificial Intelligence (AI)-Enabled Unmanned Aerial Vehicle (UAV) Systems for Optimizing User Connectivity in Sixth-Generation (6G) Ubiquitous Networks

    Zeeshan Ali Haider1, Inam Ullah2,*, Ahmad Abu Shareha3, Rashid Nasimov4, Sufyan Ali Memon5,*

    CMC-Computers, Materials & Continua, Vol.86, No.1, pp. 1-16, 2026, DOI:10.32604/cmc.2025.071042 - 10 November 2025

    Abstract The advent of sixth-generation (6G) networks introduces unprecedented challenges in achieving seamless connectivity, ultra-low latency, and efficient resource management in highly dynamic environments. Although fifth-generation (5G) networks transformed mobile broadband and machine-type communications at massive scales, their properties of scaling, interference management, and latency remain a limitation in dense high mobility settings. To overcome these limitations, artificial intelligence (AI) and unmanned aerial vehicles (UAVs) have emerged as potential solutions to develop versatile, dynamic, and energy-efficient communication systems. The study proposes an AI-based UAV architecture that utilizes cooperative reinforcement learning (CoRL) to manage an autonomous network.… More >

  • Open Access

    ARTICLE

    An Improved Reinforcement Learning-Based 6G UAV Communication for Smart Cities

    Vi Hoai Nam1, Chu Thi Minh Hue2, Dang Van Anh1,*

    CMC-Computers, Materials & Continua, Vol.86, No.1, pp. 1-15, 2026, DOI:10.32604/cmc.2025.070605 - 10 November 2025

    Abstract Unmanned Aerial Vehicles (UAVs) have become integral components in smart city infrastructures, supporting applications such as emergency response, surveillance, and data collection. However, the high mobility and dynamic topology of Flying Ad Hoc Networks (FANETs) present significant challenges for maintaining reliable, low-latency communication. Conventional geographic routing protocols often struggle in situations where link quality varies and mobility patterns are unpredictable. To overcome these limitations, this paper proposes an improved routing protocol based on reinforcement learning. This new approach integrates Q-learning with mechanisms that are both link-aware and mobility-aware. The proposed method optimizes the selection of… More >

Displaying 1-10 on page 1 of 179. Per Page