Dual-Branch Feature Continual Learning for Few-Shot Semantic Segmentation
Discuss this preprint
Start a discussion What are Sciety discussions?Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
To address the issue that large variations in appearance and scale among images of the same class lead to poor generalization of segmentation models on unseen images, a few-shot semantic segmentation method based on dual-branch feature continual learning is proposed. First, a pair of shared-weight backbone networks map the support and query images into a deep feature space. The ground truth mask of the support image is then used to separate its encoded features into foreground object features and background features. Next, the CLIP-encoded textual features are used to disentangle the mixed query features. The separated dual-branch foreground and background features are subsequently fused and aligned, enabling continual learning between branches through the target task. Finally, mask average pooling is applied to the fused foreground and background features to generate corresponding prototypes, and a parameter-free metric matching is performed to match query features with the prototype set on a per-pixel basis. Experiments on the PASCAL-5i and COCO-20i datasets under 1-shot and 5-shot settings demonstrate that the proposed method achieves competitive segmentation performance, comparable to state-of-the-art few-shot semantic segmentation methods.