Online Activation Value-aware Clustering and Aggregation for Faithful Argumentative Explanations
- Authors
- Kim, Ungsik; Bae, Jiho; Choi, Sang-Min; Lee, Suwon
- Issue Date
- Nov-2025
- Publisher
- Association for Computing Machinery, Inc
- Keywords
- aggregation function; argumentative xai; explainable ai (xai); model compression; online activation value; singular value
- Citation
- CIKM 2025 - Proceedings of the 34th ACM International Conference on Information and Knowledge Management, pp 1386 - 1395
- Pages
- 10
- Indexed
- SCOPUS
- Journal Title
- CIKM 2025 - Proceedings of the 34th ACM International Conference on Information and Knowledge Management
- Start Page
- 1386
- End Page
- 1395
- URI
- https://scholarworks.gnu.ac.kr/handle/sw.gnu/81482
- DOI
- 10.1145/3746252.3761362
- Abstract
- Argumentative explainable artificial intelligence employs argumentation theory to explain the mechanisms of machine learning. Previous approaches for explaining deep learning models collectively compressed layers via clustering. However, this resulted in accumulated information loss across layers, thereby degrading the fidelity of explanations. We propose online activation value-aware clustering and aggregation, a compression algorithm that preserves the inference structure of the original neural network with greater fidelity. The proposed method sequentially compresses each layer, immediately recalculates activation values following compression, and rectifies inter-layer information loss using a singular-value-scaled ridge alignment approach. To evaluate the effectiveness of the proposed method, we introduce four novel quantitative metrics. Input-output fidelity and structural fidelity measure how accurately the compressed model preserves the original model predictions and internal activations. Input-output perturbation consistency and structural perturbation consistency assess the similarity of the changes induced by Gaussian-perturbed input data. Experiments on three benchmark datasets (Breast Cancer, California Housing, and HIGGS) demonstrate that our method achieves performance improvements ranging from 12.9% to 53.7% across the four metrics, demonstrating significantly higher explanation fidelity than existing approaches.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - ETC > Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.