Detailed Information

Cited 2 time in webofscience Cited 3 time in scopus
Metadata Downloads

An Empirical Study on the Performance of Individual Issue Label Prediction

Authors
Heo, JueunLee, Seonah
Issue Date
Aug-2023
Publisher
Institute of Electrical and Electronics Engineers Inc.
Keywords
Empirical Study; Github; Issue Classification; Issue Report; Label Prediction; Labeling; Performance Analysis
Citation
Proceedings - 2023 IEEE/ACM 20th International Conference on Mining Software Repositories, MSR 2023, pp 228 - 233
Pages
6
Indexed
SCOPUS
Journal Title
Proceedings - 2023 IEEE/ACM 20th International Conference on Mining Software Repositories, MSR 2023
Start Page
228
End Page
233
URI
https://scholarworks.gnu.ac.kr/handle/sw.gnu/67584
DOI
10.1109/MSR59073.2023.00041
ISSN
2574-3848
2574-3864
Abstract
In GitHub, open-source software (OSS) developers label issue reports. As issue labeling is a labor-intensive manual task, automatic approaches have developed to label issue reports. However, those approaches have shown limited performance. Therefore, it is necessary to analyze the performance of predicting labels for an issue report. Understanding labels with high performance and those with low performance can help improve the performance of automatic issue labeling tasks. In this paper, we investigate the performance of individual label prediction. Our investigation uncovers labels with high performance and those with low performance. Our results can help researchers to understand the different characteristics of labels and help developers to develop a unified approach that combines several effective approaches for different kinds of issues. © 2023 IEEE.
Files in This Item
There are no files associated with this item.
Appears in
Collections
공학계열 > AI융합공학과 > Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Lee, Seon Ah photo

Lee, Seon Ah
IT공과대학 (소프트웨어공학과)
Read more

Altmetrics

Total Views & Downloads

BROWSE