Cited 3 time in
MediViSTA: Medical Video Segmentation via Temporal Fusion SAM Adaptation for Echocardiography
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Kim, Sekeun | - |
| dc.contributor.author | Jin, Pengfei | - |
| dc.contributor.author | Chen, Cheng | - |
| dc.contributor.author | Kim, Kyungsang | - |
| dc.contributor.author | Lyu, Zhiliang | - |
| dc.contributor.author | Ren, Hui | - |
| dc.contributor.author | Kim, Sunghwan | - |
| dc.contributor.author | Liu, Zhengliang | - |
| dc.contributor.author | Zhong, Aoxiao | - |
| dc.contributor.author | Liu, Tianming | - |
| dc.contributor.author | Li, Xiang | - |
| dc.contributor.author | Li, Quanzheng | - |
| dc.date.accessioned | 2025-02-26T00:30:16Z | - |
| dc.date.available | 2025-02-26T00:30:16Z | - |
| dc.date.issued | 2025 | - |
| dc.identifier.issn | 2168-2194 | - |
| dc.identifier.issn | 2168-2208 | - |
| dc.identifier.uri | https://scholarworks.gnu.ac.kr/handle/sw.gnu/77217 | - |
| dc.description.abstract | Despite achieving impressive results in general-purpose semantic segmentation with strong generalization on natural images, the Segment Anything Model (SAM) has shown less precision and stability in medical image segmentation. In particular, the SAM architecture is designed for 2D natural images and is therefore not support to handle three-dimensional information, which is particularly important for medical imaging modalities that are often volumetric or video data. In this paper, we introduce MediViSTA, a parameter-efficient fine-tuning method designed to adapt the vision foundation model for medical video, with a specific focus on echocardiography segmentation. To achieve spatial adaptation, we propose a frequency feature fusion technique that injects spatial frequency information from a CNN branch. For temporal adaptation, we integrate temporal adapters within the transformer blocks of the image encoder. Using a fine-tuning strategy, only a small subset of pre-trained parameters is updated, allowing efficient adaptation to echocardiography data. The effectiveness of our method has been comprehensively evaluated on three datasets, comprising two public datasets and one multi-center in-house dataset. Our method consistently outperforms various state-of-the-art approaches without using any prompts. Furthermore, our model exhibits strong generalization capabilities on unseen datasets, surpassing the second-best approach by 2.15% in Dice and 0.09 in temporal consistency. The results demonstrate the potential of MediViSTA to significantly advance echocardiography video segmentation, offering improved accuracy and robustness in cardiac assessment applications. © 2013 IEEE. | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | Institute of Electrical and Electronics Engineers Inc. | - |
| dc.title | MediViSTA: Medical Video Segmentation via Temporal Fusion SAM Adaptation for Echocardiography | - |
| dc.type | Article | - |
| dc.publisher.location | 미국 | - |
| dc.identifier.doi | 10.1109/JBHI.2025.3540306 | - |
| dc.identifier.scopusid | 2-s2.0-85217966407 | - |
| dc.identifier.bibliographicCitation | IEEE Journal of Biomedical and Health Informatics | - |
| dc.citation.title | IEEE Journal of Biomedical and Health Informatics | - |
| dc.type.docType | Article in press | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | scie | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.subject.keywordAuthor | Echocardiography | - |
| dc.subject.keywordAuthor | Parameter-efficient fine-tuning | - |
| dc.subject.keywordAuthor | Segment Anything Model | - |
| dc.subject.keywordAuthor | Segmentation | - |
| dc.subject.keywordAuthor | Vision Foundation model | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
Gyeongsang National University Central Library, 501, Jinju-daero, Jinju-si, Gyeongsangnam-do, 52828, Republic of Korea+82-55-772-0532
COPYRIGHT 2022 GYEONGSANG NATIONAL UNIVERSITY LIBRARY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
