Show simple item record

AuthorWang, Dan
AuthorQin, Hao
AuthorSong, Bin
AuthorXu, Ke
AuthorDu, Xiaojiang
AuthorGuizani, Mohsen
Available date2022-11-07T09:58:43Z
Publication Date2021-04-01
Publication NamePhysical Communication
Identifierhttp://dx.doi.org/10.1016/j.phycom.2020.101262
CitationWang, D., Qin, H., Song, B., Xu, K., Du, X., & Guizani, M. (2021). Joint resource allocation and power control for D2D communication with deep reinforcement learning in MCC. Physical Communication, 45, 101262.‏
ISSN18744907
URIhttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85098740731&origin=inward
URIhttp://hdl.handle.net/10576/35868
AbstractMission-critical communication (MCC) is one of the main goals in 5G, which can leverage multiple device-to-device (D2D) connections to enhance reliability for mission-critical communication. In MCC, D2D users can reuses the non-orthogonal wireless resources of cellular users without a base station (BS). Meanwhile, the D2D users will generate co-channel interference to cellular users and hence affect their quality-of-service (QoS). To comprehensively improve the user experience, we proposed a novel approach, which embraces resource allocation and power control along with Deep Reinforcement Learning (DRL). In this paper, multiple procedures are carefully designed to assist in developing our proposal. As a starter, a scenario with multiple D2D pairs and cellular users in a cell will be modeled; followed by the analysis of issues pertaining to resource allocation and power control as well as the formulation of our optimization goal; and finally, a DRL method based on spectrum allocation strategy will be created, which can ensure D2D users to obtain the sufficient resource for their QoS improvement. With the resource data provided, which D2D users capture by interacting with surroundings, the DRL method can help the D2D users autonomously selecting an available channel and power to maximize system capacity and spectrum efficiency while minimizing interference to cellular users. Experimental results show that our learning method performs well to improve resource allocation and power control significantly.
SponsorThis work has been supported by the National Natural Science Foundation of China (Nos. 61772387 , 62071354 ), the National Natural Science Foundation of Shaanxi Province, China (Grant Nos. 2019ZDLGY03-03 ), Graduate Student Innovation Fund ( 10221150004 ), and also supported by the ISN State Key Laboratory .
Languageen
PublisherElsevier B.V.
SubjectD2D
DRL
Power control
Resource allocation
TitleJoint resource allocation and power control for D2D communication with deep reinforcement learning in MCC
TypeArticle
Volume Number45


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record