An investigation on deep reinforcement learning algorithms for resource management and workload scheduling

Detalhes bibliográficos
Ano de defesa: 2022
Autor(a) principal: Abner Sousa Nascimento
Orientador(a): Não Informado pela instituição
Banca de defesa: Não Informado pela instituição
Tipo de documento: Dissertação
Tipo de acesso: Acesso aberto
Idioma: eng
Instituição de defesa: Universidade Federal de Minas Gerais
Programa de Pós-Graduação: Não Informado pela instituição
Departamento: Não Informado pela instituição
País: Não Informado pela instituição
Palavras-chave em Português:
Link de acesso: https://hdl.handle.net/1843/84048
Resumo: Efficiency is a key operational requirement for most computer systems, given that the resources necessary to such processes are usually subjected to constraints in availability. It's desirable that computing clusters operate in order to complete as many tasks as possible while making the most of hardware assets, for example, CPU and memory. In this context, the temporal ordering of the jobs submitted to a cluster can interfere in its capacity to function at maximum use. It is, thus, important that such tasks are scheduled properly to ensure efficiency. Several algorithms and techniques, both principled and learning-based, can be applied to this problem, but the goal-oriented nature of reinforcement learning powered by the use of deep neural networks can help deal with the particularities and complexities of it robustly. In this work, we investigate the usage of deep reinforcement learning techniques for job allocation in computing clusters, applying hyperparameter search and comparing the performance and training stability of the learning-based solutions with previously designed algorithms for a target metric. We found that it is possible to obtain equal or better performance under the right environmental conditions within the appropriate parametric domain. Results also indicate that such agents can achieve better generalization if trained in a graduated difficulty set-up, with increasingly challenging scenarios, instead of a random initialization approach that starts from a difficult configuration.
id UFMG_d0f4bc3275cfddf96d203d9f20a86119
oai_identifier_str oai:repositorio.ufmg.br:1843/84048
network_acronym_str UFMG
network_name_str Repositório Institucional da UFMG
repository_id_str
spelling 2025-08-05T17:08:46Z2025-09-08T22:58:49Z2025-08-05T17:08:46Z2022-12-21https://hdl.handle.net/1843/84048Efficiency is a key operational requirement for most computer systems, given that the resources necessary to such processes are usually subjected to constraints in availability. It's desirable that computing clusters operate in order to complete as many tasks as possible while making the most of hardware assets, for example, CPU and memory. In this context, the temporal ordering of the jobs submitted to a cluster can interfere in its capacity to function at maximum use. It is, thus, important that such tasks are scheduled properly to ensure efficiency. Several algorithms and techniques, both principled and learning-based, can be applied to this problem, but the goal-oriented nature of reinforcement learning powered by the use of deep neural networks can help deal with the particularities and complexities of it robustly. In this work, we investigate the usage of deep reinforcement learning techniques for job allocation in computing clusters, applying hyperparameter search and comparing the performance and training stability of the learning-based solutions with previously designed algorithms for a target metric. We found that it is possible to obtain equal or better performance under the right environmental conditions within the appropriate parametric domain. Results also indicate that such agents can achieve better generalization if trained in a graduated difficulty set-up, with increasingly challenging scenarios, instead of a random initialization approach that starts from a difficult configuration.FAPESP - Fundação de Amparo à Pesquisa do Estado de São PauloengUniversidade Federal de Minas Geraishttp://creativecommons.org/licenses/by-nd/3.0/pt/info:eu-repo/semantics/openAccessreinforcement learningdeep learninghigh power computer clustersworkload managementresource managementComputação – TesesAprendizado do computador – TesesAprendizado profundo – TesesComputação de alto desempenho – TesesAprendizado por reforço – TesesAn investigation on deep reinforcement learning algorithms for resource management and workload schedulinginfo:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/masterThesisAbner Sousa Nascimentoreponame:Repositório Institucional da UFMGinstname:Universidade Federal de Minas Gerais (UFMG)instacron:UFMGhttp://lattes.cnpq.br/6534494703690547Luis Chaimowiczhttp://lattes.cnpq.br/4499928813481251Anderson Rocha TavaresGeorge Luiz Medeiros TeodoroRenato Luiz de Freitas CunhaA eficiência é um requisito operacional fundamental para a maioria dos sistemas computacionais, visto que os recursos necessários para tais processos geralmente estão sujeitos a restrições de disponibilidade. É desejável que os clusters de computação operem para concluir o maior número possível de tarefas enquanto aproveitam ao máximo os componentes de hardware, por exemplo, CPU e memória.  Nesse contexto, a ordenação temporal das tarefas submetidas a um cluster pode interferir na sua capacidade de funcionar em uso máximo. É, portanto, importante que tais tarefas sejam programadas adequadamente para garantir a eficiência.  Vários algoritmos e técnicas, tanto orientadas por regras fundamentais quanto baseados em aprendizado de máquina, podem ser aplicados a esse problema, mas a natureza centrada em objetivos do aprendizado por reforço amplificada pelo uso de redes neurais profundas pode ajudar a lidar com as particularidades e complexidades dele de forma robusta.  Neste trabalho, investiga-se o uso de técnicas de aprendizado por reforço e redes neurais profundas para alocação de tarefas em clusters computacionais, aplicando busca pelo conjunto ideal de hiperparâmetros e comparando o desempenho e a estabilidade de treinamento das soluções baseadas em aprendizado com algoritmos previamente projetados, com referência a uma métrica alvo. Os resultados apontam que é possível obter desempenho igual ou melhor, desde que sob as condições ambientais corretas e dentro do domínio paramétrico apropriado. Observa-se também que tais agentes podem alcançar melhor generalização se treinados em uma configuração de dificuldade graduada, com cenários cada vez mais desafiadores, em vez de uma abordagem de inicialização aleatória que parte de uma configuração difícil.BrasilICX - DEPARTAMENTO DE CIÊNCIA DA COMPUTAÇÃOPrograma de Pós-Graduação em Ciência da ComputaçãoUFMGCC-LICENSElicense_rdfapplication/octet-stream805https://repositorio.ufmg.br//bitstreams/88769fef-8bed-428c-ba76-769af01bcfa9/download00e5e6a57d5512d202d12cb48704dfd6MD51falseAnonymousREADORIGINALDissertação-1.pdfapplication/pdf2614164https://repositorio.ufmg.br//bitstreams/5a4eae2e-2b04-457e-b1a0-d99dbcdbacda/downloadb49676e6de6a8cff0638373998e4e5f2MD52trueAnonymousREADLICENSElicense.txttext/plain2118https://repositorio.ufmg.br//bitstreams/2bf781d0-289c-4d9f-aa95-732ba8be6bf5/downloadcda590c95a0b51b4d15f60c9642ca272MD53falseAnonymousREAD1843/840482025-09-08 19:58:49.263http://creativecommons.org/licenses/by-nd/3.0/pt/Acesso Abertoopen.accessoai:repositorio.ufmg.br:1843/84048https://repositorio.ufmg.br/Repositório InstitucionalPUBhttps://repositorio.ufmg.br/oairepositorio@ufmg.bropendoar:2025-09-08T22:58:49Repositório Institucional da UFMG - Universidade Federal de Minas Gerais (UFMG)falseTElDRU7Dh0EgREUgRElTVFJJQlVJw4fDg08gTsODTy1FWENMVVNJVkEgRE8gUkVQT1NJVMOTUklPIElOU1RJVFVDSU9OQUwgREEgVUZNRwoKQ29tIGEgYXByZXNlbnRhw6fDo28gZGVzdGEgbGljZW7Dp2EsIHZvY8OqIChvIGF1dG9yIChlcykgb3UgbyB0aXR1bGFyIGRvcyBkaXJlaXRvcyBkZSBhdXRvcikgY29uY2VkZSBhbyBSZXBvc2l0w7NyaW8gSW5zdGl0dWNpb25hbCBkYSBVRk1HIChSSS1VRk1HKSBvIGRpcmVpdG8gbsOjbyBleGNsdXNpdm8gZSBpcnJldm9nw6F2ZWwgZGUgcmVwcm9kdXppciBlL291IGRpc3RyaWJ1aXIgYSBzdWEgcHVibGljYcOnw6NvIChpbmNsdWluZG8gbyByZXN1bW8pIHBvciB0b2RvIG8gbXVuZG8gbm8gZm9ybWF0byBpbXByZXNzbyBlIGVsZXRyw7RuaWNvIGUgZW0gcXVhbHF1ZXIgbWVpbywgaW5jbHVpbmRvIG9zIGZvcm1hdG9zIMOhdWRpbyBvdSB2w61kZW8uCgpWb2PDqiBkZWNsYXJhIHF1ZSBjb25oZWNlIGEgcG9sw610aWNhIGRlIGNvcHlyaWdodCBkYSBlZGl0b3JhIGRvIHNldSBkb2N1bWVudG8gZSBxdWUgY29uaGVjZSBlIGFjZWl0YSBhcyBEaXJldHJpemVzIGRvIFJJLVVGTUcuCgpWb2PDqiBjb25jb3JkYSBxdWUgbyBSZXBvc2l0w7NyaW8gSW5zdGl0dWNpb25hbCBkYSBVRk1HIHBvZGUsIHNlbSBhbHRlcmFyIG8gY29udGXDumRvLCB0cmFuc3BvciBhIHN1YSBwdWJsaWNhw6fDo28gcGFyYSBxdWFscXVlciBtZWlvIG91IGZvcm1hdG8gcGFyYSBmaW5zIGRlIHByZXNlcnZhw6fDo28uCgpWb2PDqiB0YW1iw6ltIGNvbmNvcmRhIHF1ZSBvIFJlcG9zaXTDs3JpbyBJbnN0aXR1Y2lvbmFsIGRhIFVGTUcgcG9kZSBtYW50ZXIgbWFpcyBkZSB1bWEgY8OzcGlhIGRlIHN1YSBwdWJsaWNhw6fDo28gcGFyYSBmaW5zIGRlIHNlZ3VyYW7Dp2EsIGJhY2stdXAgZSBwcmVzZXJ2YcOnw6NvLgoKVm9jw6ogZGVjbGFyYSBxdWUgYSBzdWEgcHVibGljYcOnw6NvIMOpIG9yaWdpbmFsIGUgcXVlIHZvY8OqIHRlbSBvIHBvZGVyIGRlIGNvbmNlZGVyIG9zIGRpcmVpdG9zIGNvbnRpZG9zIG5lc3RhIGxpY2Vuw6dhLiBWb2PDqiB0YW1iw6ltIGRlY2xhcmEgcXVlIG8gZGVww7NzaXRvIGRlIHN1YSBwdWJsaWNhw6fDo28gbsOjbywgcXVlIHNlamEgZGUgc2V1IGNvbmhlY2ltZW50bywgaW5mcmluZ2UgZGlyZWl0b3MgYXV0b3JhaXMgZGUgbmluZ3XDqW0uCgpDYXNvIGEgc3VhIHB1YmxpY2HDp8OjbyBjb250ZW5oYSBtYXRlcmlhbCBxdWUgdm9jw6ogbsOjbyBwb3NzdWkgYSB0aXR1bGFyaWRhZGUgZG9zIGRpcmVpdG9zIGF1dG9yYWlzLCB2b2PDqiBkZWNsYXJhIHF1ZSBvYnRldmUgYSBwZXJtaXNzw6NvIGlycmVzdHJpdGEgZG8gZGV0ZW50b3IgZG9zIGRpcmVpdG9zIGF1dG9yYWlzIHBhcmEgY29uY2VkZXIgYW8gUmVwb3NpdMOzcmlvIEluc3RpdHVjaW9uYWwgZGEgVUZNRyBvcyBkaXJlaXRvcyBhcHJlc2VudGFkb3MgbmVzdGEgbGljZW7Dp2EsIGUgcXVlIGVzc2UgbWF0ZXJpYWwgZGUgcHJvcHJpZWRhZGUgZGUgdGVyY2Vpcm9zIGVzdMOhIGNsYXJhbWVudGUgaWRlbnRpZmljYWRvIGUgcmVjb25oZWNpZG8gbm8gdGV4dG8gb3Ugbm8gY29udGXDumRvIGRhIHB1YmxpY2HDp8OjbyBvcmEgZGVwb3NpdGFkYS4KCkNBU08gQSBQVUJMSUNBw4fDg08gT1JBIERFUE9TSVRBREEgVEVOSEEgU0lETyBSRVNVTFRBRE8gREUgVU0gUEFUUk9Dw41OSU8gT1UgQVBPSU8gREUgVU1BIEFHw4pOQ0lBIERFIEZPTUVOVE8gT1UgT1VUUk8gT1JHQU5JU01PLCBWT0PDiiBERUNMQVJBIFFVRSBSRVNQRUlUT1UgVE9ET1MgRSBRVUFJU1FVRVIgRElSRUlUT1MgREUgUkVWSVPDg08gQ09NTyBUQU1Cw4lNIEFTIERFTUFJUyBPQlJJR0HDh8OVRVMgRVhJR0lEQVMgUE9SIENPTlRSQVRPIE9VIEFDT1JETy4KCk8gUmVwb3NpdMOzcmlvIEluc3RpdHVjaW9uYWwgZGEgVUZNRyBzZSBjb21wcm9tZXRlIGEgaWRlbnRpZmljYXIgY2xhcmFtZW50ZSBvIHNldSBub21lKHMpIG91IG8ocykgbm9tZXMocykgZG8ocykgZGV0ZW50b3IoZXMpIGRvcyBkaXJlaXRvcyBhdXRvcmFpcyBkYSBwdWJsaWNhw6fDo28sIGUgbsOjbyBmYXLDoSBxdWFscXVlciBhbHRlcmHDp8OjbywgYWzDqW0gZGFxdWVsYXMgY29uY2VkaWRhcyBwb3IgZXN0YSBsaWNlbsOnYS4K
dc.title.none.fl_str_mv An investigation on deep reinforcement learning algorithms for resource management and workload scheduling
title An investigation on deep reinforcement learning algorithms for resource management and workload scheduling
spellingShingle An investigation on deep reinforcement learning algorithms for resource management and workload scheduling
Abner Sousa Nascimento
Computação – Teses
Aprendizado do computador – Teses
Aprendizado profundo – Teses
Computação de alto desempenho – Teses
Aprendizado por reforço – Teses
reinforcement learning
deep learning
high power computer clusters
workload management
resource management
title_short An investigation on deep reinforcement learning algorithms for resource management and workload scheduling
title_full An investigation on deep reinforcement learning algorithms for resource management and workload scheduling
title_fullStr An investigation on deep reinforcement learning algorithms for resource management and workload scheduling
title_full_unstemmed An investigation on deep reinforcement learning algorithms for resource management and workload scheduling
title_sort An investigation on deep reinforcement learning algorithms for resource management and workload scheduling
author Abner Sousa Nascimento
author_facet Abner Sousa Nascimento
author_role author
dc.contributor.author.fl_str_mv Abner Sousa Nascimento
dc.subject.por.fl_str_mv Computação – Teses
Aprendizado do computador – Teses
Aprendizado profundo – Teses
Computação de alto desempenho – Teses
Aprendizado por reforço – Teses
topic Computação – Teses
Aprendizado do computador – Teses
Aprendizado profundo – Teses
Computação de alto desempenho – Teses
Aprendizado por reforço – Teses
reinforcement learning
deep learning
high power computer clusters
workload management
resource management
dc.subject.other.none.fl_str_mv reinforcement learning
deep learning
high power computer clusters
workload management
resource management
description Efficiency is a key operational requirement for most computer systems, given that the resources necessary to such processes are usually subjected to constraints in availability. It's desirable that computing clusters operate in order to complete as many tasks as possible while making the most of hardware assets, for example, CPU and memory. In this context, the temporal ordering of the jobs submitted to a cluster can interfere in its capacity to function at maximum use. It is, thus, important that such tasks are scheduled properly to ensure efficiency. Several algorithms and techniques, both principled and learning-based, can be applied to this problem, but the goal-oriented nature of reinforcement learning powered by the use of deep neural networks can help deal with the particularities and complexities of it robustly. In this work, we investigate the usage of deep reinforcement learning techniques for job allocation in computing clusters, applying hyperparameter search and comparing the performance and training stability of the learning-based solutions with previously designed algorithms for a target metric. We found that it is possible to obtain equal or better performance under the right environmental conditions within the appropriate parametric domain. Results also indicate that such agents can achieve better generalization if trained in a graduated difficulty set-up, with increasingly challenging scenarios, instead of a random initialization approach that starts from a difficult configuration.
publishDate 2022
dc.date.issued.fl_str_mv 2022-12-21
dc.date.accessioned.fl_str_mv 2025-08-05T17:08:46Z
2025-09-08T22:58:49Z
dc.date.available.fl_str_mv 2025-08-05T17:08:46Z
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/masterThesis
format masterThesis
status_str publishedVersion
dc.identifier.uri.fl_str_mv https://hdl.handle.net/1843/84048
url https://hdl.handle.net/1843/84048
dc.language.iso.fl_str_mv eng
language eng
dc.rights.driver.fl_str_mv http://creativecommons.org/licenses/by-nd/3.0/pt/
info:eu-repo/semantics/openAccess
rights_invalid_str_mv http://creativecommons.org/licenses/by-nd/3.0/pt/
eu_rights_str_mv openAccess
dc.publisher.none.fl_str_mv Universidade Federal de Minas Gerais
publisher.none.fl_str_mv Universidade Federal de Minas Gerais
dc.source.none.fl_str_mv reponame:Repositório Institucional da UFMG
instname:Universidade Federal de Minas Gerais (UFMG)
instacron:UFMG
instname_str Universidade Federal de Minas Gerais (UFMG)
instacron_str UFMG
institution UFMG
reponame_str Repositório Institucional da UFMG
collection Repositório Institucional da UFMG
bitstream.url.fl_str_mv https://repositorio.ufmg.br//bitstreams/88769fef-8bed-428c-ba76-769af01bcfa9/download
https://repositorio.ufmg.br//bitstreams/5a4eae2e-2b04-457e-b1a0-d99dbcdbacda/download
https://repositorio.ufmg.br//bitstreams/2bf781d0-289c-4d9f-aa95-732ba8be6bf5/download
bitstream.checksum.fl_str_mv 00e5e6a57d5512d202d12cb48704dfd6
b49676e6de6a8cff0638373998e4e5f2
cda590c95a0b51b4d15f60c9642ca272
bitstream.checksumAlgorithm.fl_str_mv MD5
MD5
MD5
repository.name.fl_str_mv Repositório Institucional da UFMG - Universidade Federal de Minas Gerais (UFMG)
repository.mail.fl_str_mv repositorio@ufmg.br
_version_ 1862106036004978688