TY - JOUR
T1 - Multiagent Reinforcement Learning-Based Multimodel Running Latency Optimization in Vehicular Edge Computing Paradigm
AU - Li, Peisong
AU - Xiao, Ziren
AU - Wang, Xinheng
AU - Iqbal, Muddesar
AU - Casaseca-De-La-Higuera, Pablo
N1 - Publisher Copyright:
© 2007-2012 IEEE.
PY - 2024
Y1 - 2024
N2 - With the advancement of edge computing, more and more intelligent applications are being deployed at the edge in proximity to end devices to provide in-vehicle services. However, the implementation of some complex services requires the collaboration of multiple AI models to handle and analyze various types of sensory data. In this context, the simultaneous scheduling and execution of multiple model inference tasks is an emerging scenario and faces many challenges. One of the major challenges is to reduce the completion time of time-sensitive services. In order to solve this problem, a multiagent reinforcement learning-based multimodel inference task scheduling method was proposed in this article, with a newly designed reward function to jointly optimize the overall running time and load imbalance. First, the multiagent proximal policy optimization algorithm is utilized for designing the task scheduling method. Second, the designed method can generate near-optimal task scheduling decisions and then dynamically allocate inference tasks to different edge applications based on their status and task characteristics. Third, one assessment index, quality of method, is defined and the proposed method is compared with the other five benchmark methods. Experimental results reveal that the proposed method can reduce the running time of multimodel inference by at least 25% or more, closing to the optimal solution.
AB - With the advancement of edge computing, more and more intelligent applications are being deployed at the edge in proximity to end devices to provide in-vehicle services. However, the implementation of some complex services requires the collaboration of multiple AI models to handle and analyze various types of sensory data. In this context, the simultaneous scheduling and execution of multiple model inference tasks is an emerging scenario and faces many challenges. One of the major challenges is to reduce the completion time of time-sensitive services. In order to solve this problem, a multiagent reinforcement learning-based multimodel inference task scheduling method was proposed in this article, with a newly designed reward function to jointly optimize the overall running time and load imbalance. First, the multiagent proximal policy optimization algorithm is utilized for designing the task scheduling method. Second, the designed method can generate near-optimal task scheduling decisions and then dynamically allocate inference tasks to different edge applications based on their status and task characteristics. Third, one assessment index, quality of method, is defined and the proposed method is compared with the other five benchmark methods. Experimental results reveal that the proposed method can reduce the running time of multimodel inference by at least 25% or more, closing to the optimal solution.
KW - Autonomous driving
KW - deep reinforcement learning (DRL)
KW - edge computing
KW - latency optimization
KW - multimodel inferences
KW - task scheduling
UR - http://www.scopus.com/inward/record.url?scp=85211500835&partnerID=8YFLogxK
U2 - 10.1109/JSYST.2024.3407213
DO - 10.1109/JSYST.2024.3407213
M3 - Article
AN - SCOPUS:85211500835
SN - 1932-8184
VL - 18
SP - 1860
EP - 1870
JO - IEEE Systems Journal
JF - IEEE Systems Journal
IS - 4
ER -