Show simple item record

AuthorMakhlouf, Ahmed
AuthorAbdellatif, Alaa Awad
AuthorBadawy, Ahmed
AuthorMohamed, Amr
Available date2023-11-09T06:28:03Z
Publication Date2023-01-01
Publication NameInternational Conference on Wireless and Mobile Computing, Networking and Communications
Identifierhttp://dx.doi.org/10.1109/WiMob58348.2023.10187766
CitationMakhlouf, A., Abdellatif, A. A., Badawy, A., & Mohamed, A. (2023, June). Optimized Resource and Deep Learning Model Allocation in O-RAN Architecture. In 2023 19th International Conference on Wireless and Mobile Computing, Networking and Communications (WiMob) (pp. 155-160). IEEE.‏
ISBN9798350336672
ISSN21619646
URIhttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85167621483&origin=inward
URIhttp://hdl.handle.net/10576/49122
AbstractIn the era of 5G and beyond, telecommunication networks tend to move Radio Access Network (RAN) from centralized architecture to a more distributed architecture for greater interoperability and flexibility. Open RAN (O-RAN) architecture is a paradigm shift that is proposed to enable disaggregation, virtualization, and cloudification of RAN components, possibly offered from multiple vendors, to be connected through open interfaces. Leveraging this O-RAN architecture, Deep Learning (DL) models may be running as a service close to the end users, rather than on the core network, to benefit from reduced latency and bandwidth consumption. If multiple DL models learn on the virtual edge, they will compete for the available communication and computation resources. In this paper, we introduce Optimized Resource and Model Allocation (ORMA), a framework that provides optimized resource allocation for multiple DL models learning at the edge, that aims to maximize the aggregate accuracy while respecting the limited physical resources. Distinguished from related works, ORMA optimizes the learning-related parameters, such as dataset size and number of epochs, as well as the amount of communication and computation resources allocated to each DL model to maximize the aggregate accuracy. Our results show that ORMA consistently outperforms a baseline approach that adopts a fixed, fair resource allocation (FRA) among different DL models, at different total bandwidths and CPU combinations.
Languageen
PublisherIEEE Computer Society
SubjectNetwork Slicing
Open-RAN
Optimization
Resource Allocation
Virtual Edge
TitleOptimized Resource and Deep Learning Model Allocation in O-RAN Architecture
TypeConference
Pagination155-160
Volume Number2023-June
dc.accessType Abstract Only


Files in this item

FilesSizeFormatView

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record