The
Segment Anything Model (SAM)
produces high-quality object masks from input prompts such as points or boxes, and it can be used to generate masks for all objects in an image. It has been trained on a
dataset
of 11 million images and 1.1 billion masks, and has strong zero-shot performance on a variety of segmentation tasks.
The abstract of the paper states:
We introduce the Segment Anything (SA) project: a new task, model, and dataset for image segmentation. Using our efficient model in a data collection loop, we built the largest segmentation dataset to date (by far), with over 1 billion masks on 11M licensed and privacy respecting images. The model is designed and trained to be promptable, so it can transfer zero-shot to new image distributions and tasks. We evaluate its capabilities on numerous tasks and find that its zero-shot performance is impressive -- often competitive with or even superior to prior fully supervised results. We are releasing the Segment Anything Model (SAM) and corresponding dataset (SA-1B) of 1B masks and 11M images at
https://segment-anything.com
to foster research into foundation models for computer vision.
Disclaimer
: Content from
this
model card has been written by the Hugging Face team, and parts of it were copy pasted from the original
SAM model card
.
Model Details
The SAM model is made up of 3 modules:
The
VisionEncoder
: a VIT based image encoder. It computes the image embeddings using attention on patches of the image. Relative Positional Embedding is used.
The
PromptEncoder
: generates embeddings for points and bounding boxes
The
MaskDecoder
: a two-ways transformer which performs cross attention between the image embedding and the point embeddings (->) and between the point embeddings and the image embeddings. The outputs are fed
The
Neck
: predicts the output masks based on the contextualized masks produced by the
MaskDecoder
.
If you use this model, please use the following BibTeX entry.
@article{kirillov2023segany,
title={Segment Anything},
author={Kirillov, Alexander and Mintun, Eric and Ravi, Nikhila and Mao, Hanzi and Rolland, Chloe and Gustafson, Laura and Xiao, Tete and Whitehead, Spencer and Berg, Alexander C. and Lo, Wan-Yen and Doll{\'a}r, Piotr and Girshick, Ross},
journal={arXiv:2304.02643},
year={2023}
}
Runs of wanglab medsam-vit-base on huggingface.co
1.9K
Total runs
0
24-hour runs
165
3-day runs
-71
7-day runs
-3.9K
30-day runs
More Information About medsam-vit-base huggingface.co Model
medsam-vit-base huggingface.co is an AI model on huggingface.co that provides medsam-vit-base's model effect (), which can be used instantly with this wanglab medsam-vit-base model. huggingface.co supports a free trial of the medsam-vit-base model, and also provides paid use of the medsam-vit-base. Support call medsam-vit-base model through api, including Node.js, Python, http.
medsam-vit-base huggingface.co is an online trial and call api platform, which integrates medsam-vit-base's modeling effects, including api services, and provides a free online trial of medsam-vit-base, you can try medsam-vit-base online for free by clicking the link below.
wanglab medsam-vit-base online free url in huggingface.co:
medsam-vit-base is an open source model from GitHub that offers a free installation service, and any user can find medsam-vit-base on GitHub to install. At the same time, huggingface.co provides the effect of medsam-vit-base install, users can directly use medsam-vit-base installed effect in huggingface.co for debugging and trial. It also supports api for free installation.