Next Article in Journal
Optimal Weighted Markov Model and Markov Optimal Weighted Combination Model with Their Application in Hunan’s Gross Domestic Product
Next Article in Special Issue
Dynamic Event-Triggered Sliding Mode Control of Markov Jump Delayed System with Partially Known Transition Probabilities
Previous Article in Journal
N-Bipolar Soft Expert Sets and Their Applications in Robust Multi-Attribute Group Decision-Making
Previous Article in Special Issue
Arouse-Net: Enhancing Glioblastoma Segmentation in Multi-Parametric MRI with a Custom 3D Convolutional Neural Network and Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Q-Learning Based Target Coverage Algorithm for Wireless Sensor Networks

1
Kaiserslautern Institute for Intelligent Manufacturing, Shanghai Dianji University, Shanghai 201308, China
2
School of Information Engineering, Nanchang Hangkong University, Nanchang 330063, China
3
Industrial Technology Center, Shanghai Dianji University, Shanghai 201308, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(3), 532; https://doi.org/10.3390/math13030532
Submission received: 26 November 2024 / Revised: 23 January 2025 / Accepted: 3 February 2025 / Published: 5 February 2025
(This article belongs to the Special Issue Robust Perception and Control in Prognostic Systems)

Abstract

To address the problems of unclear node activation strategy and redundant feasible solutions in solving the target coverage of wireless sensor networks, a target coverage algorithm based on deep Q-learning is proposed to learn the scheduling strategy of nodes for wireless sensor networks. First, the algorithm abstracts the construction of feasible solutions into a Markov decision process, and the smart body selects the activated sensor nodes as discrete actions according to the network environment. Second, the reward function evaluates the merit of the smart body’s choice of actions in terms of the coverage capacity of the activated nodes and their residual energy. The simulation results show that the proposed algorithm intelligences are able to stabilize their gains after 2500 rounds of learning and training under the specific designed states, actions and reward mechanisms, corresponding to the convergence of the proposed algorithm. It can also be seen that the proposed algorithm is effective under different network sizes, and its network lifetime outperforms the three greedy algorithms, the maximum lifetime coverage algorithm and the self-adaptive learning automata algorithm. Moreover, this advantage becomes more and more obvious with the increase in network size, node sensing radius and carrying initial energy.
Keywords: wireless sensor network; Q-learning; target coverage wireless sensor network; Q-learning; target coverage

Share and Cite

MDPI and ACS Style

Xiong, P.; He, D.; Lu, T. A Q-Learning Based Target Coverage Algorithm for Wireless Sensor Networks. Mathematics 2025, 13, 532. https://doi.org/10.3390/math13030532

AMA Style

Xiong P, He D, Lu T. A Q-Learning Based Target Coverage Algorithm for Wireless Sensor Networks. Mathematics. 2025; 13(3):532. https://doi.org/10.3390/math13030532

Chicago/Turabian Style

Xiong, Peng, Dan He, and Tiankun Lu. 2025. "A Q-Learning Based Target Coverage Algorithm for Wireless Sensor Networks" Mathematics 13, no. 3: 532. https://doi.org/10.3390/math13030532

APA Style

Xiong, P., He, D., & Lu, T. (2025). A Q-Learning Based Target Coverage Algorithm for Wireless Sensor Networks. Mathematics, 13(3), 532. https://doi.org/10.3390/math13030532

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop