Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs
| Ano de defesa: | 2014 |
|---|---|
| Autor(a) principal: | |
| Orientador(a): | |
| Banca de defesa: | |
| Tipo de documento: | Tese |
| Tipo de acesso: | Acesso aberto |
| Idioma: | eng |
| Instituição de defesa: |
Universidade Federal de Pernambuco
|
| Programa de Pós-Graduação: |
Não Informado pela instituição
|
| Departamento: |
Não Informado pela instituição
|
| País: |
Não Informado pela instituição
|
| Palavras-chave em Português: | |
| Link de acesso: | https://repositorio.ufpe.br/handle/123456789/12152 |
Resumo: | Currently, Object-Oriented Programming (OOP) is one of the most used paradigms. Complementarily, the software maintainability is considered a software attribute that plays an important role in quality level. In this context, the Object-Oriented Software Maintainability (OOSM) has been studied through years, and many researchers have proposed a large number of metrics to measure it. As a consequence of the number and diversity of metrics, beyond the no standardization in metrics definition and naming, the decision-making process about which metrics can be adopted in experiments on OOSM, or even their using in software companies is a difficult task. Therefore, a systematic mapping study was conducted in order to find which metrics are used as indicators in OOSM assessments. There was an initial selection of 5175 primary studies and 138 were selected, resulting in 568 metrics found. Analyzing the 568 metrics, inconsistencies in metrics’ naming were found because there were metrics with the same names but different meanings (8 cases involving 17 metrics) and also, there were metrics with different names, however with similar meanings (32 cases involving 214 metrics). Moreover, a metrics’ categorization has been proposed to facilitate decision-making process about which ones have to be adopted, and 7 categories and 17 subcategories were identified. These categories represent the evaluation scenarios where OOSM metrics should be used. Additionally, a metrics’ web portal was developed to provide information about the metrics collected in this research, and to generate metrics’ catalogs according to the context of their adoption. This portal can also be systematically fed by other researchers that work with OOSM metrics, making the results of this work the first steps towards metrics’ standardization, and the improvement of the metrics’ validation. Finally, a quasi-experiment was conducted to check the coverage index of the catalogs generated using our approach over the catalogs suggested by experts. 90% of coverage was obtained and this result was confirmed with 99% of confidential level using the Wilcoxon Test. Complementarily, a survey was conducted to check the experts’ opinion about the catalog generated by the portal when they were compared by the catalogs suggested by the experts. Thus, the coverage evaluation can be the first evidences of the usefulness of the proposed approach for metrics’ choice in OOSM evaluation. |
| id |
UFPE_2dec0b176df7750790038dd979938f32 |
|---|---|
| oai_identifier_str |
oai:repositorio.ufpe.br:123456789/12152 |
| network_acronym_str |
UFPE |
| network_name_str |
Repositório Institucional da UFPE |
| repository_id_str |
|
| spelling |
Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogsSoftware maintainabilityMetricsObject-oriented software developmentExperimental software engineeringManutenibilidade de softwareMetricasDesenvolvimento de software orientado a objetosEngenharia de software experimentalCurrently, Object-Oriented Programming (OOP) is one of the most used paradigms. Complementarily, the software maintainability is considered a software attribute that plays an important role in quality level. In this context, the Object-Oriented Software Maintainability (OOSM) has been studied through years, and many researchers have proposed a large number of metrics to measure it. As a consequence of the number and diversity of metrics, beyond the no standardization in metrics definition and naming, the decision-making process about which metrics can be adopted in experiments on OOSM, or even their using in software companies is a difficult task. Therefore, a systematic mapping study was conducted in order to find which metrics are used as indicators in OOSM assessments. There was an initial selection of 5175 primary studies and 138 were selected, resulting in 568 metrics found. Analyzing the 568 metrics, inconsistencies in metrics’ naming were found because there were metrics with the same names but different meanings (8 cases involving 17 metrics) and also, there were metrics with different names, however with similar meanings (32 cases involving 214 metrics). Moreover, a metrics’ categorization has been proposed to facilitate decision-making process about which ones have to be adopted, and 7 categories and 17 subcategories were identified. These categories represent the evaluation scenarios where OOSM metrics should be used. Additionally, a metrics’ web portal was developed to provide information about the metrics collected in this research, and to generate metrics’ catalogs according to the context of their adoption. This portal can also be systematically fed by other researchers that work with OOSM metrics, making the results of this work the first steps towards metrics’ standardization, and the improvement of the metrics’ validation. Finally, a quasi-experiment was conducted to check the coverage index of the catalogs generated using our approach over the catalogs suggested by experts. 90% of coverage was obtained and this result was confirmed with 99% of confidential level using the Wilcoxon Test. Complementarily, a survey was conducted to check the experts’ opinion about the catalog generated by the portal when they were compared by the catalogs suggested by the experts. Thus, the coverage evaluation can be the first evidences of the usefulness of the proposed approach for metrics’ choice in OOSM evaluation.Atualmente, Programao Orientada a Objetos (POO) um dos paradigmas mais utilizados. Complementarmente, a manutenibilidade de software considerada um atributo de software que desempenha um papel importante com relao ao nvel de qualidade. Neste contexto, a Manutenibilidade de Software Orientado a Objetos (MSOO) foi estudada atravs de anos e vrios pesquisadores propuseram um elevado nmero de mtricas para a medir. Como consequncia do nmero e da diversidade de mtricas existentes, alm da no padronizao nas definies e nomenclatura, a tomada de deciso sobre quais mtricas podem ser adotadas para realizar estudos em MSOO difcil. Desta forma, um mapeamento sistemtico foi realizado a fim de encontrar quais mtricas so usadas como indicadores de MSOO. Houve uma seleo inicial de 5175 estudos primrios e 138 artigos foram selecionados, resultando em 568 mtricas encontradas. Analisando as 568 mtricas, foram encontradas inconsistncias na nomenclatura destas mtricas, pois havia mtricas com nomes iguais mas significados diferentes (8 casos envolvendo 17 mtricas) e tambm mtricas com nomes diferentes e significados semelhantes (32 casos envolvendo 214 mtricas). Alm disto, uma categorizao destas mtricas foi proposta, sendo identificadas 7 categorias e 17 subcategorias. Estas categorias representam os cenrios de adoo de mtricas de MSOO. Adicionalmente, um portal web de mtricas foi desenvolvido para fornecer informaes sobre as mtricas para outros pesquisadores e tambm gerar catlogos de mtricas de acordo com o contexto da aplicao das mesmas. Este portal tambm pode ser alimentado sistematicamente por outros pesquisadores que lidam com mtricas de MSOO, fazendo com que os resultados deste trabalho possam representar os primeiros passos para padronizao e compreenso destas mtricas. Por ltimo, um quasi-experimento foi realizado para checar o grau de cobertura do catlogo proposto pela abordagem aqui apresentada quando o mesmo comparado com catlogos sugeridos por especialistas. 90% de cobertura foi obtido e este resultado foi confirmado com 99% de grau de confiana usando o Teste de Wilcoxon. De forma complementar, houve uma pesquisa de opinio para checar se os especialistas acharam catlogo gerado usando a nossa abordagem semelhante ou melhor do que o sugerido por eles. Sendo assim, os resultados da anlise da cobertura dos catlogos pode servir como indcios da utilidade da abordagem proposta para a escolha de mtricas na avaliao de MSOO.FACEPE, INESUniversidade Federal de PernambucoSOARES, Sérgio Castelo BrancoL. FILHO, Fernando J. C. deSARAIVA, Juliana de Albuquerque Gonçalves2015-03-12T14:07:25Z2015-03-12T14:07:25Z2014-01-31info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/doctoralThesisapplication/pdfhttps://repositorio.ufpe.br/handle/123456789/12152engAttribution-NonCommercial-NoDerivs 3.0 Brazilhttp://creativecommons.org/licenses/by-nc-nd/3.0/br/info:eu-repo/semantics/openAccessreponame:Repositório Institucional da UFPEinstname:Universidade Federal de Pernambuco (UFPE)instacron:UFPE2019-10-25T17:30:02Zoai:repositorio.ufpe.br:123456789/12152Repositório InstitucionalPUBhttps://repositorio.ufpe.br/oai/requestattena@ufpe.bropendoar:22212019-10-25T17:30:02Repositório Institucional da UFPE - Universidade Federal de Pernambuco (UFPE)false |
| dc.title.none.fl_str_mv |
Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs |
| title |
Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs |
| spellingShingle |
Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs SARAIVA, Juliana de Albuquerque Gonçalves Software maintainability Metrics Object-oriented software development Experimental software engineering Manutenibilidade de software Metricas Desenvolvimento de software orientado a objetos Engenharia de software experimental |
| title_short |
Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs |
| title_full |
Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs |
| title_fullStr |
Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs |
| title_full_unstemmed |
Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs |
| title_sort |
Classifying metrics for assessing object-oriented software maintainability: a family of metrics’ catalogs |
| author |
SARAIVA, Juliana de Albuquerque Gonçalves |
| author_facet |
SARAIVA, Juliana de Albuquerque Gonçalves |
| author_role |
author |
| dc.contributor.none.fl_str_mv |
SOARES, Sérgio Castelo Branco L. FILHO, Fernando J. C. de |
| dc.contributor.author.fl_str_mv |
SARAIVA, Juliana de Albuquerque Gonçalves |
| dc.subject.por.fl_str_mv |
Software maintainability Metrics Object-oriented software development Experimental software engineering Manutenibilidade de software Metricas Desenvolvimento de software orientado a objetos Engenharia de software experimental |
| topic |
Software maintainability Metrics Object-oriented software development Experimental software engineering Manutenibilidade de software Metricas Desenvolvimento de software orientado a objetos Engenharia de software experimental |
| description |
Currently, Object-Oriented Programming (OOP) is one of the most used paradigms. Complementarily, the software maintainability is considered a software attribute that plays an important role in quality level. In this context, the Object-Oriented Software Maintainability (OOSM) has been studied through years, and many researchers have proposed a large number of metrics to measure it. As a consequence of the number and diversity of metrics, beyond the no standardization in metrics definition and naming, the decision-making process about which metrics can be adopted in experiments on OOSM, or even their using in software companies is a difficult task. Therefore, a systematic mapping study was conducted in order to find which metrics are used as indicators in OOSM assessments. There was an initial selection of 5175 primary studies and 138 were selected, resulting in 568 metrics found. Analyzing the 568 metrics, inconsistencies in metrics’ naming were found because there were metrics with the same names but different meanings (8 cases involving 17 metrics) and also, there were metrics with different names, however with similar meanings (32 cases involving 214 metrics). Moreover, a metrics’ categorization has been proposed to facilitate decision-making process about which ones have to be adopted, and 7 categories and 17 subcategories were identified. These categories represent the evaluation scenarios where OOSM metrics should be used. Additionally, a metrics’ web portal was developed to provide information about the metrics collected in this research, and to generate metrics’ catalogs according to the context of their adoption. This portal can also be systematically fed by other researchers that work with OOSM metrics, making the results of this work the first steps towards metrics’ standardization, and the improvement of the metrics’ validation. Finally, a quasi-experiment was conducted to check the coverage index of the catalogs generated using our approach over the catalogs suggested by experts. 90% of coverage was obtained and this result was confirmed with 99% of confidential level using the Wilcoxon Test. Complementarily, a survey was conducted to check the experts’ opinion about the catalog generated by the portal when they were compared by the catalogs suggested by the experts. Thus, the coverage evaluation can be the first evidences of the usefulness of the proposed approach for metrics’ choice in OOSM evaluation. |
| publishDate |
2014 |
| dc.date.none.fl_str_mv |
2014-01-31 2015-03-12T14:07:25Z 2015-03-12T14:07:25Z |
| dc.type.status.fl_str_mv |
info:eu-repo/semantics/publishedVersion |
| dc.type.driver.fl_str_mv |
info:eu-repo/semantics/doctoralThesis |
| format |
doctoralThesis |
| status_str |
publishedVersion |
| dc.identifier.uri.fl_str_mv |
https://repositorio.ufpe.br/handle/123456789/12152 |
| url |
https://repositorio.ufpe.br/handle/123456789/12152 |
| dc.language.iso.fl_str_mv |
eng |
| language |
eng |
| dc.rights.driver.fl_str_mv |
Attribution-NonCommercial-NoDerivs 3.0 Brazil http://creativecommons.org/licenses/by-nc-nd/3.0/br/ info:eu-repo/semantics/openAccess |
| rights_invalid_str_mv |
Attribution-NonCommercial-NoDerivs 3.0 Brazil http://creativecommons.org/licenses/by-nc-nd/3.0/br/ |
| eu_rights_str_mv |
openAccess |
| dc.format.none.fl_str_mv |
application/pdf |
| dc.publisher.none.fl_str_mv |
Universidade Federal de Pernambuco |
| publisher.none.fl_str_mv |
Universidade Federal de Pernambuco |
| dc.source.none.fl_str_mv |
reponame:Repositório Institucional da UFPE instname:Universidade Federal de Pernambuco (UFPE) instacron:UFPE |
| instname_str |
Universidade Federal de Pernambuco (UFPE) |
| instacron_str |
UFPE |
| institution |
UFPE |
| reponame_str |
Repositório Institucional da UFPE |
| collection |
Repositório Institucional da UFPE |
| repository.name.fl_str_mv |
Repositório Institucional da UFPE - Universidade Federal de Pernambuco (UFPE) |
| repository.mail.fl_str_mv |
attena@ufpe.br |
| _version_ |
1856041927291437056 |