@inproceedings{7633caa94ea44ef5990b928edd951e7a,
title = "Analysis of Adapter in Attention of Change Detection Vision Transformer",
abstract = "Vision Transformer (ViT) contributes to accurate change detection with robustness to background changes. However, retraining ViT requires a large amount of computation to adapt to unlearned scenes. This study investigates the addition of learnable parameters into change detection ViT to reduce the computational complexity of retraining. We introduce MLP as an adapter as an addition to the attention output and the residual connection of the change detection ViT and apply LoRA method to the change detection ViT. We evaluate the retraining of additional parameter models for various background changes and analyze proper setting of additional parameters to adapt the target scenes. Introducing MLP and LoRA to change detection ViT improves the accuracy for the target scenes without competition between two additional parameter methods.",
keywords = "Adapter, Change detection, Vision Transformer",
author = "Ryunosuke Hamada and Tsubasa Minematsu and Cheng Tang and Atsushi Shimada",
note = "Publisher Copyright: {\textcopyright} The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.; 17th Asian Conference on Computer Vision, ACCV 2024 ; Conference date: 08-12-2024 Through 12-12-2024",
year = "2025",
doi = "10.1007/978-981-96-2641-0\_3",
language = "English",
isbn = "9789819626403",
series = "Lecture Notes in Computer Science",
publisher = "Springer Science and Business Media Deutschland GmbH",
pages = "36--51",
editor = "Minsu Cho and Ivan Laptev and Du Tran and Angela Yao and Hong-Bin Zha",
booktitle = "Computer Vision – ACCV 2024 Workshops - 17th Asian Conference on Computer Vision, Revised Selected Papers",
address = "Germany",
}