Cited 1 time in
Deep Generative Replay with Denoising Diffusion Probabilistic Models for Continual Learning in Audio Classification
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Lee, Hyeon-Ju | - |
| dc.contributor.author | Buu, Seok-Jun | - |
| dc.date.accessioned | 2024-12-03T05:30:40Z | - |
| dc.date.available | 2024-12-03T05:30:40Z | - |
| dc.date.issued | 2024-09 | - |
| dc.identifier.issn | 2169-3536 | - |
| dc.identifier.issn | 2169-3536 | - |
| dc.identifier.uri | https://scholarworks.gnu.ac.kr/handle/sw.gnu/74333 | - |
| dc.description.abstract | Accurate classification of audio data is essential in various fields such as speech recognition, safety management, healthcare, security, and surveillance. However, existing deep learning classifiers typically require extensive pre-collected data and struggle to adapt to the emergence of new audio classes over time. To address these challenges, this paper proposes a continual learning method utilizing Diffusion-driven Generative Replay (DDGR). The proposed DDGR method continuously updates the model at each training stage with high-quality generated data from Denoising Diffusion Probabilistic Models (DDPM), preserving existing knowledge. Furthermore, by embedding disentangled representations through a triplet network, the model can effectively recognize new classes as they emerge. This approach overcomes the problem of catastrophic forgetting and effectively resolves the issue of data scalability in a continual learning setup. The proposed method achieved the highest AIA values of 95.45% and 72.99% on the Audio MNIST and ESC-50 datasets, respectively, compared to existing continual learning methods. Additionally, for Audio MNIST, it showed IM-0.01, FWT 0.27, FM 0.06, and BWT-0.06, indicating that it best preserves prior knowledge while learning new data most effectively. For ESC-50, it demonstrated IM of-0.12, FWT of 0.09, FM of 0.17, and BWT of-0.17. These results validate the efficacy of the DDGR method in maintaining prior knowledge while integrating new information and highlight the complementary role of the triplet network in enhancing feature representation. © 2013 IEEE. | - |
| dc.format.extent | 14 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | Institute of Electrical and Electronics Engineers Inc. | - |
| dc.title | Deep Generative Replay with Denoising Diffusion Probabilistic Models for Continual Learning in Audio Classification | - |
| dc.type | Article | - |
| dc.publisher.location | 미국 | - |
| dc.identifier.doi | 10.1109/ACCESS.2024.3459954 | - |
| dc.identifier.scopusid | 2-s2.0-85204088440 | - |
| dc.identifier.wosid | 001327334100001 | - |
| dc.identifier.bibliographicCitation | IEEE Access, v.12, pp 134714 - 134727 | - |
| dc.citation.title | IEEE Access | - |
| dc.citation.volume | 12 | - |
| dc.citation.startPage | 134714 | - |
| dc.citation.endPage | 134727 | - |
| dc.type.docType | Article | - |
| dc.description.isOpenAccess | Y | - |
| dc.description.journalRegisteredClass | scie | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.relation.journalResearchArea | Computer Science | - |
| dc.relation.journalResearchArea | Engineering | - |
| dc.relation.journalResearchArea | Telecommunications | - |
| dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
| dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
| dc.relation.journalWebOfScienceCategory | Telecommunications | - |
| dc.subject.keywordAuthor | Audio classification | - |
| dc.subject.keywordAuthor | Continual learning | - |
| dc.subject.keywordAuthor | Denoising diffusion probabilistic model | - |
| dc.subject.keywordAuthor | Generative replay | - |
| dc.subject.keywordAuthor | Triplet network | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
Gyeongsang National University Central Library, 501, Jinju-daero, Jinju-si, Gyeongsangnam-do, 52828, Republic of Korea+82-55-772-0532
COPYRIGHT 2022 GYEONGSANG NATIONAL UNIVERSITY LIBRARY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
