Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs

Detalhes bibliográficos
Ano de defesa: 2014
Autor(a) principal: SARAIVA, Juliana de Albuquerque Gonçalves
Orientador(a): Não Informado pela instituição
Banca de defesa: Não Informado pela instituição
Tipo de documento: Tese
Tipo de acesso: Acesso aberto
Idioma: eng
Instituição de defesa: Universidade Federal de Pernambuco
Programa de Pós-Graduação: Não Informado pela instituição
Departamento: Não Informado pela instituição
País: Não Informado pela instituição
Palavras-chave em Português:
Link de acesso: https://repositorio.ufpe.br/handle/123456789/12152
Resumo: Currently, Object-Oriented Programming (OOP) is one of the most used paradigms. Complementarily, the software maintainability is considered a software attribute that plays an important role in quality level. In this context, the Object-Oriented Software Maintainability (OOSM) has been studied through years, and many researchers have proposed a large number of metrics to measure it. As a consequence of the number and diversity of metrics, beyond the no standardization in metrics definition and naming, the decision-making process about which metrics can be adopted in experiments on OOSM, or even their using in software companies is a difficult task. Therefore, a systematic mapping study was conducted in order to find which metrics are used as indicators in OOSM assessments. There was an initial selection of 5175 primary studies and 138 were selected, resulting in 568 metrics found. Analyzing the 568 metrics, inconsistencies in metrics’ naming were found because there were metrics with the same names but different meanings (8 cases involving 17 metrics) and also, there were metrics with different names, however with similar meanings (32 cases involving 214 metrics). Moreover, a metrics’ categorization has been proposed to facilitate decision-making process about which ones have to be adopted, and 7 categories and 17 subcategories were identified. These categories represent the evaluation scenarios where OOSM metrics should be used. Additionally, a metrics’ web portal was developed to provide information about the metrics collected in this research, and to generate metrics’ catalogs according to the context of their adoption. This portal can also be systematically fed by other researchers that work with OOSM metrics, making the results of this work the first steps towards metrics’ standardization, and the improvement of the metrics’ validation. Finally, a quasi-experiment was conducted to check the coverage index of the catalogs generated using our approach over the catalogs suggested by experts. 90% of coverage was obtained and this result was confirmed with 99% of confidential level using the Wilcoxon Test. Complementarily, a survey was conducted to check the experts’ opinion about the catalog generated by the portal when they were compared by the catalogs suggested by the experts. Thus, the coverage evaluation can be the first evidences of the usefulness of the proposed approach for metrics’ choice in OOSM evaluation.
id UFPE_2dec0b176df7750790038dd979938f32
oai_identifier_str oai:repositorio.ufpe.br:123456789/12152
network_acronym_str UFPE
network_name_str Repositório Institucional da UFPE
repository_id_str
spelling Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogsSoftware maintainabilityMetricsObject-oriented software developmentExperimental software engineeringManutenibilidade de softwareMetricasDesenvolvimento de software orientado a objetosEngenharia de software experimentalCurrently, Object-Oriented Programming (OOP) is one of the most used paradigms. Complementarily, the software maintainability is considered a software attribute that plays an important role in quality level. In this context, the Object-Oriented Software Maintainability (OOSM) has been studied through years, and many researchers have proposed a large number of metrics to measure it. As a consequence of the number and diversity of metrics, beyond the no standardization in metrics definition and naming, the decision-making process about which metrics can be adopted in experiments on OOSM, or even their using in software companies is a difficult task. Therefore, a systematic mapping study was conducted in order to find which metrics are used as indicators in OOSM assessments. There was an initial selection of 5175 primary studies and 138 were selected, resulting in 568 metrics found. Analyzing the 568 metrics, inconsistencies in metrics’ naming were found because there were metrics with the same names but different meanings (8 cases involving 17 metrics) and also, there were metrics with different names, however with similar meanings (32 cases involving 214 metrics). Moreover, a metrics’ categorization has been proposed to facilitate decision-making process about which ones have to be adopted, and 7 categories and 17 subcategories were identified. These categories represent the evaluation scenarios where OOSM metrics should be used. Additionally, a metrics’ web portal was developed to provide information about the metrics collected in this research, and to generate metrics’ catalogs according to the context of their adoption. This portal can also be systematically fed by other researchers that work with OOSM metrics, making the results of this work the first steps towards metrics’ standardization, and the improvement of the metrics’ validation. Finally, a quasi-experiment was conducted to check the coverage index of the catalogs generated using our approach over the catalogs suggested by experts. 90% of coverage was obtained and this result was confirmed with 99% of confidential level using the Wilcoxon Test. Complementarily, a survey was conducted to check the experts’ opinion about the catalog generated by the portal when they were compared by the catalogs suggested by the experts. Thus, the coverage evaluation can be the first evidences of the usefulness of the proposed approach for metrics’ choice in OOSM evaluation.Atualmente, Programao Orientada a Objetos (POO) um dos paradigmas mais utilizados. Complementarmente, a manutenibilidade de software considerada um atributo de software que desempenha um papel importante com relao ao nvel de qualidade. Neste contexto, a Manutenibilidade de Software Orientado a Objetos (MSOO) foi estudada atravs de anos e vrios pesquisadores propuseram um elevado nmero de mtricas para a medir. Como consequncia do nmero e da diversidade de mtricas existentes, alm da no padronizao nas definies e nomenclatura, a tomada de deciso sobre quais mtricas podem ser adotadas para realizar estudos em MSOO difcil. Desta forma, um mapeamento sistemtico foi realizado a fim de encontrar quais mtricas so usadas como indicadores de MSOO. Houve uma seleo inicial de 5175 estudos primrios e 138 artigos foram selecionados, resultando em 568 mtricas encontradas. Analisando as 568 mtricas, foram encontradas inconsistncias na nomenclatura destas mtricas, pois havia mtricas com nomes iguais mas significados diferentes (8 casos envolvendo 17 mtricas) e tambm mtricas com nomes diferentes e significados semelhantes (32 casos envolvendo 214 mtricas). Alm disto, uma categorizao destas mtricas foi proposta, sendo identificadas 7 categorias e 17 subcategorias. Estas categorias representam os cenrios de adoo de mtricas de MSOO. Adicionalmente, um portal web de mtricas foi desenvolvido para fornecer informaes sobre as mtricas para outros pesquisadores e tambm gerar catlogos de mtricas de acordo com o contexto da aplicao das mesmas. Este portal tambm pode ser alimentado sistematicamente por outros pesquisadores que lidam com mtricas de MSOO, fazendo com que os resultados deste trabalho possam representar os primeiros passos para padronizao e compreenso destas mtricas. Por ltimo, um quasi-experimento foi realizado para checar o grau de cobertura do catlogo proposto pela abordagem aqui apresentada quando o mesmo comparado com catlogos sugeridos por especialistas. 90% de cobertura foi obtido e este resultado foi confirmado com 99% de grau de confiana usando o Teste de Wilcoxon. De forma complementar, houve uma pesquisa de opinio para checar se os especialistas acharam catlogo gerado usando a nossa abordagem semelhante ou melhor do que o sugerido por eles. Sendo assim, os resultados da anlise da cobertura dos catlogos pode servir como indcios da utilidade da abordagem proposta para a escolha de mtricas na avaliao de MSOO.FACEPE, INESUniversidade Federal de PernambucoSOARES, Sérgio Castelo BrancoL. FILHO, Fernando J. C. deSARAIVA, Juliana de Albuquerque Gonçalves2015-03-12T14:07:25Z2015-03-12T14:07:25Z2014-01-31info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/doctoralThesisapplication/pdfhttps://repositorio.ufpe.br/handle/123456789/12152engAttribution-NonCommercial-NoDerivs 3.0 Brazilhttp://creativecommons.org/licenses/by-nc-nd/3.0/br/info:eu-repo/semantics/openAccessreponame:Repositório Institucional da UFPEinstname:Universidade Federal de Pernambuco (UFPE)instacron:UFPE2019-10-25T17:30:02Zoai:repositorio.ufpe.br:123456789/12152Repositório InstitucionalPUBhttps://repositorio.ufpe.br/oai/requestattena@ufpe.bropendoar:22212019-10-25T17:30:02Repositório Institucional da UFPE - Universidade Federal de Pernambuco (UFPE)false
dc.title.none.fl_str_mv Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs
title Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs
spellingShingle Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs
SARAIVA, Juliana de Albuquerque Gonçalves
Software maintainability
Metrics
Object-oriented software development
Experimental software engineering
Manutenibilidade de software
Metricas
Desenvolvimento de software orientado a objetos
Engenharia de software experimental
title_short Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs
title_full Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs
title_fullStr Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs
title_full_unstemmed Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs
title_sort Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs
author SARAIVA, Juliana de Albuquerque Gonçalves
author_facet SARAIVA, Juliana de Albuquerque Gonçalves
author_role author
dc.contributor.none.fl_str_mv SOARES, Sérgio Castelo Branco
L. FILHO, Fernando J. C. de
dc.contributor.author.fl_str_mv SARAIVA, Juliana de Albuquerque Gonçalves
dc.subject.por.fl_str_mv Software maintainability
Metrics
Object-oriented software development
Experimental software engineering
Manutenibilidade de software
Metricas
Desenvolvimento de software orientado a objetos
Engenharia de software experimental
topic Software maintainability
Metrics
Object-oriented software development
Experimental software engineering
Manutenibilidade de software
Metricas
Desenvolvimento de software orientado a objetos
Engenharia de software experimental
description Currently, Object-Oriented Programming (OOP) is one of the most used paradigms. Complementarily, the software maintainability is considered a software attribute that plays an important role in quality level. In this context, the Object-Oriented Software Maintainability (OOSM) has been studied through years, and many researchers have proposed a large number of metrics to measure it. As a consequence of the number and diversity of metrics, beyond the no standardization in metrics definition and naming, the decision-making process about which metrics can be adopted in experiments on OOSM, or even their using in software companies is a difficult task. Therefore, a systematic mapping study was conducted in order to find which metrics are used as indicators in OOSM assessments. There was an initial selection of 5175 primary studies and 138 were selected, resulting in 568 metrics found. Analyzing the 568 metrics, inconsistencies in metrics’ naming were found because there were metrics with the same names but different meanings (8 cases involving 17 metrics) and also, there were metrics with different names, however with similar meanings (32 cases involving 214 metrics). Moreover, a metrics’ categorization has been proposed to facilitate decision-making process about which ones have to be adopted, and 7 categories and 17 subcategories were identified. These categories represent the evaluation scenarios where OOSM metrics should be used. Additionally, a metrics’ web portal was developed to provide information about the metrics collected in this research, and to generate metrics’ catalogs according to the context of their adoption. This portal can also be systematically fed by other researchers that work with OOSM metrics, making the results of this work the first steps towards metrics’ standardization, and the improvement of the metrics’ validation. Finally, a quasi-experiment was conducted to check the coverage index of the catalogs generated using our approach over the catalogs suggested by experts. 90% of coverage was obtained and this result was confirmed with 99% of confidential level using the Wilcoxon Test. Complementarily, a survey was conducted to check the experts’ opinion about the catalog generated by the portal when they were compared by the catalogs suggested by the experts. Thus, the coverage evaluation can be the first evidences of the usefulness of the proposed approach for metrics’ choice in OOSM evaluation.
publishDate 2014
dc.date.none.fl_str_mv 2014-01-31
2015-03-12T14:07:25Z
2015-03-12T14:07:25Z
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/doctoralThesis
format doctoralThesis
status_str publishedVersion
dc.identifier.uri.fl_str_mv https://repositorio.ufpe.br/handle/123456789/12152
url https://repositorio.ufpe.br/handle/123456789/12152
dc.language.iso.fl_str_mv eng
language eng
dc.rights.driver.fl_str_mv Attribution-NonCommercial-NoDerivs 3.0 Brazil
http://creativecommons.org/licenses/by-nc-nd/3.0/br/
info:eu-repo/semantics/openAccess
rights_invalid_str_mv Attribution-NonCommercial-NoDerivs 3.0 Brazil
http://creativecommons.org/licenses/by-nc-nd/3.0/br/
eu_rights_str_mv openAccess
dc.format.none.fl_str_mv application/pdf
dc.publisher.none.fl_str_mv Universidade Federal de Pernambuco
publisher.none.fl_str_mv Universidade Federal de Pernambuco
dc.source.none.fl_str_mv reponame:Repositório Institucional da UFPE
instname:Universidade Federal de Pernambuco (UFPE)
instacron:UFPE
instname_str Universidade Federal de Pernambuco (UFPE)
instacron_str UFPE
institution UFPE
reponame_str Repositório Institucional da UFPE
collection Repositório Institucional da UFPE
repository.name.fl_str_mv Repositório Institucional da UFPE - Universidade Federal de Pernambuco (UFPE)
repository.mail.fl_str_mv attena@ufpe.br
_version_ 1856041927291437056