A Comparison of Pretrained Models for Classifying Issue Reportsopen access
- Authors
- Heo, Jueun; Kwon, Gibeom; Kwak, Changwon; Lee, Seonah
- Issue Date
- Jun-2024
- Publisher
- Institute of Electrical and Electronics Engineers Inc.
- Keywords
- BERT; Bidirectional control; Codes; Computer bugs; Data models; deep learning techniques; Encoding; issue classification; issue reports; pretrained models; Software engineering; Task analysis
- Citation
- IEEE Access, v.12, pp 1 - 1
- Pages
- 1
- Indexed
- SCIE
SCOPUS
- Journal Title
- IEEE Access
- Volume
- 12
- Start Page
- 1
- End Page
- 1
- URI
- https://scholarworks.gnu.ac.kr/handle/sw.gnu/70859
- DOI
- 10.1109/ACCESS.2024.3408688
- ISSN
- 2169-3536
2169-3536
- Abstract
- Issues are evolving requirements which are the main factor that increase the cost of software evolution. To help developers manage issues, GitHub provides issue labeling mechanisms in issue management systems. However, manually labeling issue reports still requires considerable developer effort. To ease developers’ burden, researchers have proposed automatically classifying issue reports. They used deep learning techniques and pretrained models to improve the classification accuracy. However, pretrained models in the general domain such as RoBERTa have limitations in understanding the contexts of software engineering tasks. In this paper, we create a pretrained model IssueBERT with issue data to understand if a domain-specific pretrained model could improve the accuracy of classifying issue reports. We also adopt and explore several pretrained models with software engineering domains, CodeBERT, BERTOverflow, and seBERT. We conduct a comparative experiment of these pretrained models to understand their performance in classifying issue reports. Our comparison results show that IssueBERT outperforms other pretrained models. Noticeably, IssueBERT yields an average F1 score that is 1.74% higher than that of seBERT and 3.61% higher than that of RoBERTa, even though IssueBERT was pretrained with much less data than seBERT and RoBERTa. Authors
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - 공학계열 > AI융합공학과 > Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.