@@ -120,11 +120,11 @@ def _cfg(url='', **kwargs):
120120 input_size = (3 , 384 , 384 ), crop_pct = 1.0 , classifier = ('head' , 'head_dist' )),
121121
122122 # ViT ImageNet-21K-P pretraining
123- 'vit_base_patch16_224_in21k_miil ' : _cfg (
123+ 'vit_base_patch16_224_miil_in21k ' : _cfg (
124124 url = 'https://miil-public-eu.oss-eu-central-1.aliyuncs.com/model-zoo/ImageNet_21K_P/models/timm/vit_base_patch16_224_in21k_miil.pth' ,
125125 mean = (0 , 0 , 0 ), std = (1 , 1 , 1 ), crop_pct = 0.875 , interpolation = 'bilinear' , num_classes = 11221 ,
126126 ),
127- 'vit_base_patch16_224_1k_miil ' : _cfg (
127+ 'vit_base_patch16_224_miil ' : _cfg (
128128 url = 'https://miil-public-eu.oss-eu-central-1.aliyuncs.com/model-zoo/ImageNet_21K_P/models/timm'
129129 '/vit_base_patch16_224_1k_miil_84_4.pth' ,
130130 mean = (0 , 0 , 0 ), std = (1 , 1 , 1 ), crop_pct = 0.875 , interpolation = 'bilinear' ,
@@ -699,20 +699,22 @@ def vit_deit_base_distilled_patch16_384(pretrained=False, **kwargs):
699699 'vit_deit_base_distilled_patch16_384' , pretrained = pretrained , distilled = True , ** model_kwargs )
700700 return model
701701
702+
702703@register_model
703- def vit_base_patch16_224_in21k_miil (pretrained = False , ** kwargs ):
704+ def vit_base_patch16_224_miil_in21k (pretrained = False , ** kwargs ):
704705 """ ViT-Base (ViT-B/16) from original paper (https://arxiv.org/abs/2010.11929).
705706 Weights taken from: https://github.com/Alibaba-MIIL/ImageNet21K
706707 """
707708 model_kwargs = dict (patch_size = 16 , embed_dim = 768 , depth = 12 , num_heads = 12 , qkv_bias = False , ** kwargs )
708- model = _create_vision_transformer ('vit_base_patch16_224_in21k_miil ' , pretrained = pretrained , ** model_kwargs )
709+ model = _create_vision_transformer ('vit_base_patch16_224_miil_in21k ' , pretrained = pretrained , ** model_kwargs )
709710 return model
710711
712+
711713@register_model
712- def vit_base_patch16_224_1k_miil (pretrained = False , ** kwargs ):
714+ def vit_base_patch16_224_miil (pretrained = False , ** kwargs ):
713715 """ ViT-Base (ViT-B/16) from original paper (https://arxiv.org/abs/2010.11929).
714716 Weights taken from: https://github.com/Alibaba-MIIL/ImageNet21K
715717 """
716718 model_kwargs = dict (patch_size = 16 , embed_dim = 768 , depth = 12 , num_heads = 12 , qkv_bias = False , ** kwargs )
717- model = _create_vision_transformer ('vit_base_patch16_224_1k_miil ' , pretrained = pretrained , ** model_kwargs )
719+ model = _create_vision_transformer ('vit_base_patch16_224_miil ' , pretrained = pretrained , ** model_kwargs )
718720 return model
0 commit comments