Skip to content

Latest commit

 

History

History
163 lines (119 loc) · 8.9 KB

File metadata and controls

163 lines (119 loc) · 8.9 KB

Efficient-Segment-Anything-Model

Segment Anything Model (SAM) has attracted considerable attention from the community since its inception. However, the formidable model size and demanding computational requirements of SAM have rendered it cumbersome for deployment on resource-constrained devices. To mitigate these constraints, many efforts have been made to effectively make SAM more efficient and lightweight. This repository provides a summary of those efficient segment anything models.

If you find this repository helpful, please consider Stars ⭐ or Sharing ⬆️. Thanks.

Updates

  • 🚀 August 7, 2024: Updated new related works before 07/08/2024 in this GitHub.
  • 🚀 March 22, 2024: Updated new related works before 20/03/2024 in this GitHub.

Contents

Scratch Training Methods

FastSAM

Fast Segment Anything
Xu Zhao, Wenchao Ding, Yongqi An, Yinglong Du, Tao Yu, Min Li, Ming Tang, Jinqiao Wang
Institute of Automation, Chinese Academy of Sciences, Beijing, China
Paper: [Arxiv] Code: Stars

Description: The Fast Segment Anything Model(FastSAM) is a CNN Segment Anything Model trained using only 2% of the SA-1B dataset published by SAM authors. FastSAM

Lite-SAM

Lite-SAM Is Actually What You Need for Segment Everything
Jianhai Fu, Yuanjie Yu, Ningchuan Li, Yi Zhang, Qichao Chen, Jianping Xiong, Jun Yin, and Zhiyu Xiang
Zhejiang Dahua Technology, Zhejiang University, Hangzhou, China
Paper: [Arxiv]

Description: Lite-SAM is an efficient end-to-end solution for the SegEvery task designed to reduce computational costs and redundancy. litesam

Knowledge Distillation Methods

MobileSAM

FASTER SEGMENT ANYTHING: TOWARDS LIGHTWEIGHT SAM FOR MOBILE APPLICATIONS
Chaoning Zhang, Dongshen Han, Sheng Zheng, Jinwoo Choi, Tae-Ho Kim, Choong Seon Hong
Kyung Hee University
Paper: [Arxiv] Code: Stars

Description: MobileSAM is a lightweight SAM suitable for resource-constrained mobile devices. MobileSAM

EdgeSAM

EdgeSAM: Prompt-In-the-Loop Distillation for On-Device Deployment of SAM
Chong Zhou, Xiangtai Li, Chen Change Loy, Bo Dai
S-Lab, Nanyang Technological University, Shanghai Artificial Intelligence Laborator
Paper: [Arxiv] Code: Stars

Description: EdgeSAM is an accelerated variant of the Segment Anything Model (SAM), optimized for efficient execution on edge devices with minimal compromise in performance. MobileSAM

EfficientSAM

EfficientSAM: Leveraged Masked Image Pretraining for Efficient Segment Anything
Yunyang Xiong et al.
Meta AI Research
Paper: [Arxiv] Code: Stars

Description: EfficientSAMs is lightweight SAM model that exhibits decent performance with largely reduced complexity. The idea is based on leveraging masked image pretraining, SAMI, which learns to reconstruct features from SAM image encoder for effective visual representation learning. EfficientSAM

EfficientViT-SAM

EfficientViT-SAM: Accelerated Segment Anything Model Without Performance Loss
Zhuoyang Zhang, Han Cai, Song Han
Tsinghua University, MIT, NVIDIA
Paper: [Arxiv] Code: Stars

Description: EfficientViT-SAM is a new family of accelerated segment anything models. It retain SAM’s lightweight prompt encoder and mask decoder while replacing the heavy image encoder with EfficientViT EfficientViT-SAM

RepViT-SAM

RepViT-SAM: Towards Real-Time Segmenting Anythings
Ao Wang, Hui Chen, Zijia Lin, Jungong Han, Guiguang Ding
Tsinghua University, The University of Sheffield
Paper: [Arxiv] Code: Stars

Description: RepViT achieves the state-of-the-art performance and latency trade-off on mobile devices by incorporating efficient architectural designs of ViTs into CNNs. Here, to achieve real-time segmenting anything on mobile devices, following [27], this work replace the heavyweight image encoder in SAM with RepViT model, ending up with the RepViT-SAM model. RepViT-SAM

SAM-LIGHTENING

SAM-LIGHTENING: A LIGHTWEIGHT SEGMENT ANYTHING MODEL WITHIN DILATED FLASH ATTENTION TO ACHIEVE 30× ACCELERATION
Yanfei Song, Bangzheng Pu, Peng Wang, Dong Dong, Hongxu Jiang, Yiqing Shen
Beihang University, Beihang Hangzhou Innovation Institute, Johns Hopkins University
Paper: [Arxiv] Code: [Anonymous Github]

Description: SAM-Lightening revolutionizes image segmentation by introducing a highly efficient, lightweight model. This model features an innovative Dilated Flash Attention mechanism, enabling rapid inference and minimal memory usage, ideal for applications in real-time environments and resource-constrained devices. SAM-LIGHTENING

TinySAM

TinySAM: Pushing the Envelope for Efficient Segment Anything Model
Han Shu, Wenshuo Li, Yehui Tang, Yiman Zhang, Yihao Chen, Houqiang Li, Yunhe Wang, Xinghao Che
University of Science and Technology of China, Huawei Noah’s Ark Lab
Paper: [Arxiv] Code: Stars

Description: TinySAM is a framework can obtain a tiny segment anything model while maintaining the strong zero-shot performance. TinySAM

Model Pruning Methods

SlimSAM

SlimSAM: 0.1% Data Makes Segment Anything Slim
Zigeng Chen, Gongfan Fang, Xinyin Ma, Xinchao Wang
Learning and Vision Lab, National University of Singapore
Paper: [Arxiv] Code: Stars

Description: SlimSAM is a novel data-efficient SAM compression method that achieves superior performance with extremely less training data. The essence of SlimSAM is encapsulated in the alternate slimming framework which effectively enhances knowledge inheritance under severely limited training data availability and exceptional pruning ratio.

SlimSAM

Training Free Methods

Expedit-SAM

Expediting Large-Scale Vision Transformer for Dense Prediction without Fine-tuning
Weicong Liang et al.
Key Laboratory of Machine Perception, Peking University, ETH Zurich
Paper: [Arxiv] Code: Stars

Description: Expedit-SAM is a method can speed up SAM without any training. This is achieved by a novel token clustering algorithm. SAM-LIGHTENING1 SAM-LIGHTENING2