Options
Trautsch, Alexander Richard
Loading...
Preferred name
Trautsch, Alexander Richard
Official Name
Trautsch, Alexander Richard
Alternative Name
Trautsch, Alexander
Trautsch, A.
Trautsch, A. R.
Main Affiliation
ORCID
Now showing 1 - 2 of 2
2022Journal Article [["dc.bibliographiccitation.issue","6"],["dc.bibliographiccitation.journal","Empirical Software Engineering"],["dc.bibliographiccitation.volume","27"],["dc.contributor.author","Herbold, Steffen"],["dc.contributor.author","Trautsch, Alexander"],["dc.contributor.author","Ledel, Benjamin"],["dc.contributor.author","Aghamohammadi, Alireza"],["dc.contributor.author","Ghaleb, Taher A."],["dc.contributor.author","Chahal, Kuljit Kaur"],["dc.contributor.author","Bossenmaier, Tim"],["dc.contributor.author","Nagaria, Bhaveet"],["dc.contributor.author","Makedonski, Philip"],["dc.contributor.author","Ahmadabadi, Matin Nili"],["dc.contributor.author","Erbel, Johannes"],["dc.date.accessioned","2022-09-01T09:49:20Z"],["dc.date.available","2022-09-01T09:49:20Z"],["dc.date.issued","2022"],["dc.description.abstract","Abstract\n \n Context\n Tangled commits are changes to software that address multiple concerns at once. For researchers interested in bugs, tangled commits mean that they actually study not only bugs, but also other concerns irrelevant for the study of bugs.\n \n \n Objective\n We want to improve our understanding of the prevalence of tangling and the types of changes that are tangled within bug fixing commits.\n \n \n Methods\n We use a crowd sourcing approach for manual labeling to validate which changes contribute to bug fixes for each line in bug fixing commits. Each line is labeled by four participants. If at least three participants agree on the same label, we have consensus.\n \n \n Results\n We estimate that between 17% and 32% of all changes in bug fixing commits modify the source code to fix the underlying problem. However, when we only consider changes to the production code files this ratio increases to 66% to 87%. We find that about 11% of lines are hard to label leading to active disagreements between participants. Due to confirmed tangling and the uncertainty in our data, we estimate that 3% to 47% of data is noisy without manual untangling, depending on the use case.\n \n \n Conclusion\n Tangled commits have a high prevalence in bug fixes and can lead to a large amount of noise in the data. Prior research indicates that this noise may alter results. As researchers, we should be skeptics and assume that unvalidated data is likely very noisy, until proven otherwise."],["dc.identifier.doi","10.1007/s10664-021-10083-5"],["dc.identifier.pii","10083"],["dc.identifier.uri","https://resolver.sub.uni-goettingen.de/purl?gro-2/113398"],["dc.language.iso","en"],["dc.notes.intern","DOI-Import GROB-597"],["dc.relation.eissn","1573-7616"],["dc.relation.issn","1382-3256"],["dc.rights.uri","https://creativecommons.org/licenses/by/4.0"],["dc.title","A fine-grained data set and analysis of tangling in bug fixing commits"],["dc.type","journal_article"],["dc.type.internalPublication","yes"],["dspace.entity.type","Publication"]]Details DOI2022Journal Article [["dc.bibliographiccitation.artnumber","42"],["dc.bibliographiccitation.issue","2"],["dc.bibliographiccitation.journal","Empirical Software Engineering"],["dc.bibliographiccitation.volume","27"],["dc.contributor.author","Herbold, Steffen"],["dc.contributor.author","Trautsch, Alexander"],["dc.contributor.author","Trautsch, Fabian"],["dc.contributor.author","Ledel, Benjamin"],["dc.date.accessioned","2022-02-01T10:31:56Z"],["dc.date.available","2022-02-01T10:31:56Z"],["dc.date.issued","2022"],["dc.description.abstract","Abstract Context The SZZ algorithm is the de facto standard for labeling bug fixing commits and finding inducing changes for defect prediction data. Recent research uncovered potential problems in different parts of the SZZ algorithm. Most defect prediction data sets provide only static code metrics as features, while research indicates that other features are also important. Objective We provide an empirical analysis of the defect labels created with the SZZ algorithm and the impact of commonly used features on results. Method We used a combination of manual validation and adopted or improved heuristics for the collection of defect data. We conducted an empirical study on 398 releases of 38 Apache projects. Results We found that only half of the bug fixing commits determined by SZZ are actually bug fixing. If a six-month time frame is used in combination with SZZ to determine which bugs affect a release, one file is incorrectly labeled as defective for every file that is correctly labeled as defective. In addition, two defective files are missed. We also explored the impact of the relatively small set of features that are available in most defect prediction data sets, as there are multiple publications that indicate that, e.g., churn related features are important for defect prediction. We found that the difference of using more features is not significant. Conclusion Problems with inaccurate defect labels are a severe threat to the validity of the state of the art of defect prediction. Small feature sets seem to be a less severe threat."],["dc.description.abstract","Abstract Context The SZZ algorithm is the de facto standard for labeling bug fixing commits and finding inducing changes for defect prediction data. Recent research uncovered potential problems in different parts of the SZZ algorithm. Most defect prediction data sets provide only static code metrics as features, while research indicates that other features are also important. Objective We provide an empirical analysis of the defect labels created with the SZZ algorithm and the impact of commonly used features on results. Method We used a combination of manual validation and adopted or improved heuristics for the collection of defect data. We conducted an empirical study on 398 releases of 38 Apache projects. Results We found that only half of the bug fixing commits determined by SZZ are actually bug fixing. If a six-month time frame is used in combination with SZZ to determine which bugs affect a release, one file is incorrectly labeled as defective for every file that is correctly labeled as defective. In addition, two defective files are missed. We also explored the impact of the relatively small set of features that are available in most defect prediction data sets, as there are multiple publications that indicate that, e.g., churn related features are important for defect prediction. We found that the difference of using more features is not significant. Conclusion Problems with inaccurate defect labels are a severe threat to the validity of the state of the art of defect prediction. Small feature sets seem to be a less severe threat."],["dc.identifier.doi","10.1007/s10664-021-10092-4"],["dc.identifier.pii","10092"],["dc.identifier.uri","https://resolver.sub.uni-goettingen.de/purl?gro-2/98985"],["dc.language.iso","en"],["dc.notes.intern","DOI-Import GROB-517"],["dc.relation.eissn","1573-7616"],["dc.relation.issn","1382-3256"],["dc.rights.uri","https://creativecommons.org/licenses/by/4.0"],["dc.title","Problems with SZZ and features: An empirical study of the state of practice of defect prediction data collection"],["dc.type","journal_article"],["dc.type.internalPublication","yes"],["dspace.entity.type","Publication"]]Details DOI