A Study on Applying Large Language Models to Issue Classification
- Authors
- Heo, Jueun; Lee, Seonah
- Issue Date
- Jun-2025
- Keywords
- Classification; GPT-3.5; GPT-4; Instruction Tuning; Issue; Llama 3.1 8B; OpenAI
- Citation
- IEEE International Conference on Program Comprehension, pp 136 - 146
- Pages
- 11
- Indexed
- SCOPUS
- Journal Title
- IEEE International Conference on Program Comprehension
- Start Page
- 136
- End Page
- 146
- URI
- https://scholarworks.gnu.ac.kr/handle/sw.gnu/79411
- DOI
- 10.1109/ICPC66645.2025.00022
- ISSN
- 2643-7147
2643-7171
- Abstract
- Prompt-based large language models (LLMs) have demonstrated their ability to perform tasks with minimal or no additional training data. In the context of issue classification, researchers have actively explored the capabilities of LLMs in classifying issue reports. However, existing studies still face limitations in accuracy. This study replicates an LLM-based issue classification study using GPT-3.5 Turbo and explores variants, such as adopting different models like Llama 3.18 B and GPT-4o. Experimental results show that the classifier fine-tuned with GPT-3.5 Turbo still yields the same accuracy as shown in the original research and that the classifier fine-tuned with Llama 3.18 B(0.8004) yields an F1-score of 0.0535 lower than that of the classifier fine-tuned with GPT-3.5 Turbo (0.8467). On the other hand, the classifier with GPT-4o (0.8639) yields an average F1-score 0. 01 higher than that of the classifier fine-tuned with GPT-3.5 Turbo (0.8467). Additionally, the project-agnostic classifier fine-tuned with GPT-4o yields the highest F1-score of 0.8680. These findings contribute to advancing LLM-based issue classification by providing experimental insights into the accuracy of LLMs in this issue classification task. © 2025 IEEE.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - 공학계열 > AI융합공학과 > Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.