跳到主要导航 跳到搜索 跳到主要内容

M-DRL: Deep Reinforcement Learning Based Coflow Traffic Scheduler with MLFQ Threshold Adaption

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

The coflow scheduling in data-parallel clusters can improve application-level communication performance. The existing coflow scheduling method without prior knowledge usually uses Multi-Level Feedback Queue (MLFQ) with fixed threshold parameters, which is insensitive to coflow traffic characteristics. Manual adjustment of the threshold parameters for different application scenarios often has long optimization period and is coarse in optimization granularity. We propose M-DRL, a deep reinforcement learning based coflow traffic scheduler by dynamically setting thresholds of MLFQ to adapt to the coflow traffic characteristics, and reduces the average coflow completion time. Trace-driven simulations on the public dataset show that coflow communication stages using M-DRL complete 2.08 × (6.48 × ) and 1.36 × (1.25 × ) faster on average coflow completion time (95-th percentile) in comparison to per-flow fairness and Aalo, and is comparable to SEBF with prior knowledge.

源语言英语
主期刊名Network and Parallel Computing - 17th IFIP WG 10.3 International Conference, NPC 2020, Revised Selected Papers
编辑Xin He, En Shao, Guangming Tan
出版商Springer Science and Business Media Deutschland GmbH
80-91
页数12
ISBN(印刷版)9783030794774
DOI
出版状态已出版 - 2021
活动17th IFIP WG 10.3 International Conference on Network and Parallel Computing, NPC 2020 - Zhengzhou, 中国
期限: 28 9月 202030 9月 2020

出版系列

姓名Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
12639 LNCS
ISSN(印刷版)0302-9743
ISSN(电子版)1611-3349

会议

会议17th IFIP WG 10.3 International Conference on Network and Parallel Computing, NPC 2020
国家/地区中国
Zhengzhou
时期28/09/2030/09/20

指纹

探究 'M-DRL: Deep Reinforcement Learning Based Coflow Traffic Scheduler with MLFQ Threshold Adaption' 的科研主题。它们共同构成独一无二的指纹。

引用此