Back to Search Start Over

kNN-CLIP: Retrieval Enables Training-Free Segmentation on Continually Expanding Large Vocabularies

Authors :
Gui, Zhongrui
Sun, Shuyang
Li, Runjia
Yuan, Jianhao
An, Zhaochong
Roth, Karsten
Prabhu, Ameya
Torr, Philip
Gui, Zhongrui
Sun, Shuyang
Li, Runjia
Yuan, Jianhao
An, Zhaochong
Roth, Karsten
Prabhu, Ameya
Torr, Philip
Publication Year :
2024

Abstract

Rapid advancements in continual segmentation have yet to bridge the gap of scaling to large continually expanding vocabularies under compute-constrained scenarios. We discover that traditional continual training leads to catastrophic forgetting under compute constraints, unable to outperform zero-shot segmentation methods. We introduce a novel strategy for semantic and panoptic segmentation with zero forgetting, capable of adapting to continually growing vocabularies without the need for retraining or large memory costs. Our training-free approach, kNN-CLIP, leverages a database of instance embeddings to enable open-vocabulary segmentation approaches to continually expand their vocabulary on any given domain with a single-pass through data, while only storing embeddings minimizing both compute and memory costs. This method achieves state-of-the-art mIoU performance across large-vocabulary semantic and panoptic segmentation datasets. We hope kNN-CLIP represents a step forward in enabling more efficient and adaptable continual segmentation, paving the way for advances in real-world large-vocabulary continual segmentation methods.<br />Comment: 10 pages, 3 figures

Details

Database :
OAIster
Publication Type :
Electronic Resource
Accession number :
edsoai.on1438546731
Document Type :
Electronic Resource