from urllib.request import urlopen
from PIL import Image
import timm
img = Image.open(urlopen(
'https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/beignets-task-guide.png'
))
model = timm.create_model('swinv2_large_window12to16_192to256.ms_in22k_ft_in1k', pretrained=True)
model = model.eval()
# get model specific transforms (normalization, resize)
data_config = timm.data.resolve_model_data_config(model)
transforms = timm.data.create_transform(**data_config, is_training=False)
output = model(transforms(img).unsqueeze(0)) # unsqueeze single image into batch of 1
top5_probabilities, top5_class_indices = torch.topk(output.softmax(dim=1) * 100, k=5)
Feature Map Extraction
from urllib.request import urlopen
from PIL import Image
import timm
img = Image.open(urlopen(
'https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/beignets-task-guide.png'
))
model = timm.create_model(
'swinv2_large_window12to16_192to256.ms_in22k_ft_in1k',
pretrained=True,
features_only=True,
)
model = model.eval()
# get model specific transforms (normalization, resize)
data_config = timm.data.resolve_model_data_config(model)
transforms = timm.data.create_transform(**data_config, is_training=False)
output = model(transforms(img).unsqueeze(0)) # unsqueeze single image into batch of 1for o in output:
# print shape of each feature map in output# e.g. for swin_base_patch4_window7_224 (NHWC output)# torch.Size([1, 56, 56, 128])# torch.Size([1, 28, 28, 256])# torch.Size([1, 14, 14, 512])# torch.Size([1, 7, 7, 1024])# e.g. for swinv2_cr_small_ns_224 (NCHW output)# torch.Size([1, 96, 56, 56]) # torch.Size([1, 192, 28, 28])# torch.Size([1, 384, 14, 14])# torch.Size([1, 768, 7, 7])print(o.shape)
Image Embeddings
from urllib.request import urlopen
from PIL import Image
import timm
img = Image.open(urlopen(
'https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/beignets-task-guide.png'
))
model = timm.create_model(
'swinv2_large_window12to16_192to256.ms_in22k_ft_in1k',
pretrained=True,
num_classes=0, # remove classifier nn.Linear
)
model = model.eval()
# get model specific transforms (normalization, resize)
data_config = timm.data.resolve_model_data_config(model)
transforms = timm.data.create_transform(**data_config, is_training=False)
output = model(transforms(img).unsqueeze(0)) # output is (batch_size, num_features) shaped tensor# or equivalently (without needing to set num_classes=0)
output = model.forward_features(transforms(img).unsqueeze(0))
# output is unpooled (ie.e a (batch_size, H, W, num_features) tensor for swin / swinv2# or (batch_size, num_features, H, W) for swinv2_cr
output = model.forward_head(output, pre_logits=True)
# output is (batch_size, num_features) tensor
Model Comparison
Explore the dataset and runtime metrics of this model in timm
model results
.
Citation
@inproceedings{liu2021swinv2,
title={Swin Transformer V2: Scaling Up Capacity and Resolution},
author={Ze Liu and Han Hu and Yutong Lin and Zhuliang Yao and Zhenda Xie and Yixuan Wei and Jia Ning and Yue Cao and Zheng Zhang and Li Dong and Furu Wei and Baining Guo},
booktitle={International Conference on Computer Vision and Pattern Recognition (CVPR)},
year={2022}
}
@misc{rw2019timm,
author = {Ross Wightman},
title = {PyTorch Image Models},
year = {2019},
publisher = {GitHub},
journal = {GitHub repository},
doi = {10.5281/zenodo.4414861},
howpublished = {\url{https://github.com/huggingface/pytorch-image-models}}
}
Runs of timm swinv2_large_window12to16_192to256.ms_in22k_ft_in1k on huggingface.co
123.7K
Total runs
0
24-hour runs
-25.1K
3-day runs
-71.1K
7-day runs
49.5K
30-day runs
More Information About swinv2_large_window12to16_192to256.ms_in22k_ft_in1k huggingface.co Model
More swinv2_large_window12to16_192to256.ms_in22k_ft_in1k license Visit here:
swinv2_large_window12to16_192to256.ms_in22k_ft_in1k huggingface.co is an AI model on huggingface.co that provides swinv2_large_window12to16_192to256.ms_in22k_ft_in1k's model effect (), which can be used instantly with this timm swinv2_large_window12to16_192to256.ms_in22k_ft_in1k model. huggingface.co supports a free trial of the swinv2_large_window12to16_192to256.ms_in22k_ft_in1k model, and also provides paid use of the swinv2_large_window12to16_192to256.ms_in22k_ft_in1k. Support call swinv2_large_window12to16_192to256.ms_in22k_ft_in1k model through api, including Node.js, Python, http.
swinv2_large_window12to16_192to256.ms_in22k_ft_in1k huggingface.co is an online trial and call api platform, which integrates swinv2_large_window12to16_192to256.ms_in22k_ft_in1k's modeling effects, including api services, and provides a free online trial of swinv2_large_window12to16_192to256.ms_in22k_ft_in1k, you can try swinv2_large_window12to16_192to256.ms_in22k_ft_in1k online for free by clicking the link below.
timm swinv2_large_window12to16_192to256.ms_in22k_ft_in1k online free url in huggingface.co:
swinv2_large_window12to16_192to256.ms_in22k_ft_in1k is an open source model from GitHub that offers a free installation service, and any user can find swinv2_large_window12to16_192to256.ms_in22k_ft_in1k on GitHub to install. At the same time, huggingface.co provides the effect of swinv2_large_window12to16_192to256.ms_in22k_ft_in1k install, users can directly use swinv2_large_window12to16_192to256.ms_in22k_ft_in1k installed effect in huggingface.co for debugging and trial. It also supports api for free installation.
swinv2_large_window12to16_192to256.ms_in22k_ft_in1k install url in huggingface.co: