Disclaimer: The team releasing Mask2Former did not write a model card for this model so this model card has been written by the Hugging Face team.
Model description
Mask2Former addresses instance, semantic and panoptic segmentation with the same paradigm: by predicting a set of masks and corresponding labels. Hence, all 3 tasks are treated as if they were instance segmentation. Mask2Former outperforms the previous SOTA,
MaskFormer
both in terms of performance an efficiency by (i) replacing the pixel decoder with a more advanced multi-scale deformable attention Transformer, (ii) adopting a Transformer decoder with masked attention to boost performance without
without introducing additional computation and (iii) improving training efficiency by calculating the loss on subsampled points instead of whole masks.
Intended uses & limitations
You can use this particular checkpoint for instance segmentation. See the
model hub
to look for other
fine-tuned versions on a task that interests you.
How to use
Here is how to use this model:
import requests
import torch
from PIL import Image
from transformers import AutoImageProcessor, Mask2FormerForUniversalSegmentation
# load Mask2Former fine-tuned on COCO instance segmentation
processor = AutoImageProcessor.from_pretrained("facebook/mask2former-swin-large-coco-instance")
model = Mask2FormerForUniversalSegmentation.from_pretrained("facebook/mask2former-swin-large-coco-instance")
url = "http://images.cocodataset.org/val2017/000000039769.jpg"
image = Image.open(requests.get(url, stream=True).raw)
inputs = processor(images=image, return_tensors="pt")
with torch.no_grad():
outputs = model(**inputs)
# model predicts class_queries_logits of shape `(batch_size, num_queries)`# and masks_queries_logits of shape `(batch_size, num_queries, height, width)`
class_queries_logits = outputs.class_queries_logits
masks_queries_logits = outputs.masks_queries_logits
# you can pass them to processor for postprocessing
result = processor.post_process_instance_segmentation(outputs, target_sizes=[image.size[::-1]])[0]
# we refer to the demo notebooks for visualization (see "Resources" section in the Mask2Former docs)
predicted_instance_map = result["segmentation"]
mask2former-swin-large-coco-instance huggingface.co is an AI model on huggingface.co that provides mask2former-swin-large-coco-instance's model effect (), which can be used instantly with this facebook mask2former-swin-large-coco-instance model. huggingface.co supports a free trial of the mask2former-swin-large-coco-instance model, and also provides paid use of the mask2former-swin-large-coco-instance. Support call mask2former-swin-large-coco-instance model through api, including Node.js, Python, http.
mask2former-swin-large-coco-instance huggingface.co is an online trial and call api platform, which integrates mask2former-swin-large-coco-instance's modeling effects, including api services, and provides a free online trial of mask2former-swin-large-coco-instance, you can try mask2former-swin-large-coco-instance online for free by clicking the link below.
facebook mask2former-swin-large-coco-instance online free url in huggingface.co:
mask2former-swin-large-coco-instance is an open source model from GitHub that offers a free installation service, and any user can find mask2former-swin-large-coco-instance on GitHub to install. At the same time, huggingface.co provides the effect of mask2former-swin-large-coco-instance install, users can directly use mask2former-swin-large-coco-instance installed effect in huggingface.co for debugging and trial. It also supports api for free installation.
mask2former-swin-large-coco-instance install url in huggingface.co: