About | Contact Us | Register | Login
ProceedingsSeriesJournalsSearchEAI
Machine Learning and Intelligent Communication. 8th EAI International Conference, MLICOM 2023, Beijing, China, December 17, 2023, Proceedings

Research Article

Self-guided Few-Shot Semantic Segmentation for Remote Sensing Imagery Based on Large Vision Models

Cite
BibTeX Plain Text
  • @INPROCEEDINGS{10.1007/978-3-031-71716-1_6,
        author={Xiyu Qi and Yifan Wu and Yongqiang Mao and Wenhui Zhang and Yidan Zhang},
        title={Self-guided Few-Shot Semantic Segmentation for Remote Sensing Imagery Based on Large Vision Models},
        proceedings={Machine Learning and Intelligent Communication. 8th EAI International Conference, MLICOM 2023, Beijing, China, December 17, 2023, Proceedings},
        proceedings_a={MLICOM},
        year={2024},
        month={9},
        keywords={Remote sensing images visual foundation model semantic segmentation prompt learning},
        doi={10.1007/978-3-031-71716-1_6}
    }
    
  • Xiyu Qi
    Yifan Wu
    Yongqiang Mao
    Wenhui Zhang
    Yidan Zhang
    Year: 2024
    Self-guided Few-Shot Semantic Segmentation for Remote Sensing Imagery Based on Large Vision Models
    MLICOM
    Springer
    DOI: 10.1007/978-3-031-71716-1_6
Xiyu Qi1, Yifan Wu1, Yongqiang Mao1, Wenhui Zhang1, Yidan Zhang1,*
  • 1: Aerospace Information Research Institute, Chinese Academy of Sciences
*Contact email: zhangyd@aircas.ac.cn

Abstract

The Segment Anything Model (SAM) exhibits remarkable versatility and zero-shot learning abilities, owing largely to its extensive training data (SA-1B). Recognizing SAM’s dependency on manual guidance given its category-agnostic nature, we identified unexplored potential within few-shot semantic segmentation tasks for remote sensing imagery. This research introduces a structured framework designed for the automation of few-shot semantic segmentation. It utilizes the SAM model and facilitates a more efficient generation of semantically discernible segmentation outcomes. Central to our methodology is a novel automatic prompt learning approach, leveraging prior guided mask to produce coarse pixel-wise prompts for SAM. Extensive experiments on the DLRSD datasets underlines the superiority of our approach, outperforming other available few-shot methodologies.

Keywords
Remote sensing images visual foundation model semantic segmentation prompt learning
Published
2024-09-20
Appears in
SpringerLink
http://dx.doi.org/10.1007/978-3-031-71716-1_6
Copyright © 2023–2025 ICST
EBSCOProQuestDBLPDOAJPortico
EAI Logo

About EAI

  • Who We Are
  • Leadership
  • Research Areas
  • Partners
  • Media Center

Community

  • Membership
  • Conference
  • Recognition
  • Sponsor Us

Publish with EAI

  • Publishing
  • Journals
  • Proceedings
  • Books
  • EUDL