Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquina

Detalhes bibliográficos
Ano de defesa: 2022
Autor(a) principal: Samara Silva Santos
Orientador(a): Não Informado pela instituição
Banca de defesa: Não Informado pela instituição
Tipo de documento: Dissertação
Tipo de acesso: Acesso aberto
Idioma: por
Instituição de defesa: Universidade Federal de Minas Gerais
Programa de Pós-Graduação: Não Informado pela instituição
Departamento: Não Informado pela instituição
País: Não Informado pela instituição
Palavras-chave em Português:
Link de acesso: https://hdl.handle.net/1843/47408
Resumo: Machine Learning methods (ML) have been widely used in several applications, due to the high power of generalization and the ability to obtain complex relationships between data. Although systems achieve this feat, there is usually no clear relationship as to why a particular decision was made, as well as the impact of changing attributes on the generated outputs. The need to understand these methods becomes even more present in the face of laws that guarantee the ``right to explanation'', as provided for in article 20 of the General Data Protection Law (LGPD), and in other regulations in this sense throughout the world. As a result, this work proposes to investigate the induction of Oblique Decision Trees - also known as Perceptron Decision Tree or PDT - as a method of local interpretability for complex ML models. Since the PDT is transparent, it can be used to locally simulate the behavior of more complex models and thus extract information about them through it. With this in mind, a local approximation of the predictions of the complex method to be explained was proposed, through the induction of PDTs, whose weights evolved through a heuristic optimization technique, based on evolutionary computation. With the grown tree, explanations about the local decisions of opaque models are generated, by providing the rules followed to obtain the outputs, exposing the hierarchy of local importance of the attributes and decision limits associated with each one of them. A new PDT model for regression problems was also presented, which is used to generate local explanations for this type of problem. The final application generated was named Perceptron Decision Tree Explainer (or PDTX), which, in short, is a model-agnostic local interpretability method, which works with structured tabular data, and which can make a better approximation than some classical methods in the literature, maintaining, in addition to the stability of the generated explanations, their simplicity. Additionally, a study was made on the effect of applying three local sampling techniques together with PDTX, concerning the stability of the generated explanations, and the reduction of dimensionality by five methods of reduction of attributes present in the literature, on the impact of the quality of the local approach. The results obtained are promising: compared to LIME (Local Interpretable Model-Agnostic Explanations) and Decision Trees (DT), PDTX performed significantly better for known metrics such as fidelity and stability, both in the context of classification, as in regression, and is comparable to LIME in terms of simplicity.
id UFMG_eea42ceb37fac57e4628727dbf0c020b
oai_identifier_str oai:repositorio.ufmg.br:1843/47408
network_acronym_str UFMG
network_name_str Repositório Institucional da UFMG
repository_id_str
spelling Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquinaEngenharia elétricaAprendizado do computadorÁrvores de decisãoInteligência artificialInteligência artificial explicávelInterpretabilidade em IAInteligência artificialAprendizado de máquinaÁrvores de decisões oblíquasMachine Learning methods (ML) have been widely used in several applications, due to the high power of generalization and the ability to obtain complex relationships between data. Although systems achieve this feat, there is usually no clear relationship as to why a particular decision was made, as well as the impact of changing attributes on the generated outputs. The need to understand these methods becomes even more present in the face of laws that guarantee the ``right to explanation'', as provided for in article 20 of the General Data Protection Law (LGPD), and in other regulations in this sense throughout the world. As a result, this work proposes to investigate the induction of Oblique Decision Trees - also known as Perceptron Decision Tree or PDT - as a method of local interpretability for complex ML models. Since the PDT is transparent, it can be used to locally simulate the behavior of more complex models and thus extract information about them through it. With this in mind, a local approximation of the predictions of the complex method to be explained was proposed, through the induction of PDTs, whose weights evolved through a heuristic optimization technique, based on evolutionary computation. With the grown tree, explanations about the local decisions of opaque models are generated, by providing the rules followed to obtain the outputs, exposing the hierarchy of local importance of the attributes and decision limits associated with each one of them. A new PDT model for regression problems was also presented, which is used to generate local explanations for this type of problem. The final application generated was named Perceptron Decision Tree Explainer (or PDTX), which, in short, is a model-agnostic local interpretability method, which works with structured tabular data, and which can make a better approximation than some classical methods in the literature, maintaining, in addition to the stability of the generated explanations, their simplicity. Additionally, a study was made on the effect of applying three local sampling techniques together with PDTX, concerning the stability of the generated explanations, and the reduction of dimensionality by five methods of reduction of attributes present in the literature, on the impact of the quality of the local approach. The results obtained are promising: compared to LIME (Local Interpretable Model-Agnostic Explanations) and Decision Trees (DT), PDTX performed significantly better for known metrics such as fidelity and stability, both in the context of classification, as in regression, and is comparable to LIME in terms of simplicity.CNPq - Conselho Nacional de Desenvolvimento Científico e TecnológicoFAPEMIG - Fundação de Amparo à Pesquisa do Estado de Minas GeraisCAPES - Coordenação de Aperfeiçoamento de Pessoal de Nível SuperiorUniversidade Federal de Minas Gerais2022-11-23T18:35:28Z2025-09-08T23:47:05Z2022-11-23T18:35:28Z2022-07-14info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/masterThesisapplication/pdfhttps://hdl.handle.net/1843/47408porSamara Silva Santosinfo:eu-repo/semantics/openAccessreponame:Repositório Institucional da UFMGinstname:Universidade Federal de Minas Gerais (UFMG)instacron:UFMG2025-09-08T23:47:05Zoai:repositorio.ufmg.br:1843/47408Repositório InstitucionalPUBhttps://repositorio.ufmg.br/oairepositorio@ufmg.bropendoar:2025-09-08T23:47:05Repositório Institucional da UFMG - Universidade Federal de Minas Gerais (UFMG)false
dc.title.none.fl_str_mv Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquina
title Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquina
spellingShingle Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquina
Samara Silva Santos
Engenharia elétrica
Aprendizado do computador
Árvores de decisão
Inteligência artificial
Inteligência artificial explicável
Interpretabilidade em IA
Inteligência artificial
Aprendizado de máquina
Árvores de decisões oblíquas
title_short Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquina
title_full Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquina
title_fullStr Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquina
title_full_unstemmed Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquina
title_sort Indução de árvores de decisão oblíquas como explicadores de predições por modelos de aprendizado de máquina
author Samara Silva Santos
author_facet Samara Silva Santos
author_role author
dc.contributor.author.fl_str_mv Samara Silva Santos
dc.subject.por.fl_str_mv Engenharia elétrica
Aprendizado do computador
Árvores de decisão
Inteligência artificial
Inteligência artificial explicável
Interpretabilidade em IA
Inteligência artificial
Aprendizado de máquina
Árvores de decisões oblíquas
topic Engenharia elétrica
Aprendizado do computador
Árvores de decisão
Inteligência artificial
Inteligência artificial explicável
Interpretabilidade em IA
Inteligência artificial
Aprendizado de máquina
Árvores de decisões oblíquas
description Machine Learning methods (ML) have been widely used in several applications, due to the high power of generalization and the ability to obtain complex relationships between data. Although systems achieve this feat, there is usually no clear relationship as to why a particular decision was made, as well as the impact of changing attributes on the generated outputs. The need to understand these methods becomes even more present in the face of laws that guarantee the ``right to explanation'', as provided for in article 20 of the General Data Protection Law (LGPD), and in other regulations in this sense throughout the world. As a result, this work proposes to investigate the induction of Oblique Decision Trees - also known as Perceptron Decision Tree or PDT - as a method of local interpretability for complex ML models. Since the PDT is transparent, it can be used to locally simulate the behavior of more complex models and thus extract information about them through it. With this in mind, a local approximation of the predictions of the complex method to be explained was proposed, through the induction of PDTs, whose weights evolved through a heuristic optimization technique, based on evolutionary computation. With the grown tree, explanations about the local decisions of opaque models are generated, by providing the rules followed to obtain the outputs, exposing the hierarchy of local importance of the attributes and decision limits associated with each one of them. A new PDT model for regression problems was also presented, which is used to generate local explanations for this type of problem. The final application generated was named Perceptron Decision Tree Explainer (or PDTX), which, in short, is a model-agnostic local interpretability method, which works with structured tabular data, and which can make a better approximation than some classical methods in the literature, maintaining, in addition to the stability of the generated explanations, their simplicity. Additionally, a study was made on the effect of applying three local sampling techniques together with PDTX, concerning the stability of the generated explanations, and the reduction of dimensionality by five methods of reduction of attributes present in the literature, on the impact of the quality of the local approach. The results obtained are promising: compared to LIME (Local Interpretable Model-Agnostic Explanations) and Decision Trees (DT), PDTX performed significantly better for known metrics such as fidelity and stability, both in the context of classification, as in regression, and is comparable to LIME in terms of simplicity.
publishDate 2022
dc.date.none.fl_str_mv 2022-11-23T18:35:28Z
2022-11-23T18:35:28Z
2022-07-14
2025-09-08T23:47:05Z
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/masterThesis
format masterThesis
status_str publishedVersion
dc.identifier.uri.fl_str_mv https://hdl.handle.net/1843/47408
url https://hdl.handle.net/1843/47408
dc.language.iso.fl_str_mv por
language por
dc.rights.driver.fl_str_mv info:eu-repo/semantics/openAccess
eu_rights_str_mv openAccess
dc.format.none.fl_str_mv application/pdf
dc.publisher.none.fl_str_mv Universidade Federal de Minas Gerais
publisher.none.fl_str_mv Universidade Federal de Minas Gerais
dc.source.none.fl_str_mv reponame:Repositório Institucional da UFMG
instname:Universidade Federal de Minas Gerais (UFMG)
instacron:UFMG
instname_str Universidade Federal de Minas Gerais (UFMG)
instacron_str UFMG
institution UFMG
reponame_str Repositório Institucional da UFMG
collection Repositório Institucional da UFMG
repository.name.fl_str_mv Repositório Institucional da UFMG - Universidade Federal de Minas Gerais (UFMG)
repository.mail.fl_str_mv repositorio@ufmg.br
_version_ 1856413995074846720