Skip to main navigation Skip to search Skip to main content

Hierarchical Reinforcement Learning with Self-Distillation for Resource Scheduling in Complex Resource Networks

  • Kexin Zhang*
  • , Qing Gao
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

The resource scheduling problem still presents various limitations in traditional methods when dealing with a large number of resource packages. This paper proposes a novel hierarchical reinforcement learning with self-distillation (HRLSD) model to address the resource scheduling problem in complex service networks with a large-scale of resource packages. To reduce the complexity of the policy space, we introduce a student-teacher encoder as a hierarchical reinforcement learning model. The student Q-value and teacher Q-value are obtained from two separate Q-networks using the deep Q-learning method. By distilling the knowledge from the teacher Q-value to the student Q-value, the student encoder learns from the teacher encoder to enhance effectiveness without increasing computational complexity. We evaluate the effectiveness of our proposed method through numerical examples involving large-scale resource packages.

Original languageEnglish
Title of host publicationISCAS 2025 - IEEE International Symposium on Circuits and Systems, Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350356830
DOIs
StatePublished - 2025
Event2025 IEEE International Symposium on Circuits and Systems, ISCAS 2025 - London, United Kingdom
Duration: 25 May 202528 May 2025

Publication series

NameProceedings - IEEE International Symposium on Circuits and Systems
ISSN (Print)0271-4310

Conference

Conference2025 IEEE International Symposium on Circuits and Systems, ISCAS 2025
Country/TerritoryUnited Kingdom
CityLondon
Period25/05/2528/05/25

Keywords

  • complex service networks
  • hierarchical reinforcement learning (HRL)
  • resource scheduling
  • self-distillation

Fingerprint

Dive into the research topics of 'Hierarchical Reinforcement Learning with Self-Distillation for Resource Scheduling in Complex Resource Networks'. Together they form a unique fingerprint.

Cite this