@@ -477,6 +477,11 @@ def _cfg(url='', **kwargs):
477477 hf_hub_id = 'timm/' ,
478478 mean = IMAGENET_DEFAULT_MEAN , std = IMAGENET_DEFAULT_STD
479479 ),
480+ 'beitv2_base_patch16_224.in1k_ft_in1k' : _cfg (
481+ url = 'https://conversationhub.blob.core.windows.net/beit-share-public/beitv2/beitv2_base_patch16_224_pt1k_ft1k.pth' ,
482+ hf_hub_id = 'timm/' ,
483+ mean = IMAGENET_DEFAULT_MEAN , std = IMAGENET_DEFAULT_STD
484+ ),
480485 'beitv2_base_patch16_224.in1k_ft_in22k' : _cfg (
481486 url = 'https://conversationhub.blob.core.windows.net/beit-share-public/beitv2/beitv2_base_patch16_224_pt1k_ft21k.pth' ,
482487 hf_hub_id = 'timm/' ,
@@ -487,6 +492,11 @@ def _cfg(url='', **kwargs):
487492 hf_hub_id = 'timm/' ,
488493 crop_pct = 0.95 , mean = IMAGENET_DEFAULT_MEAN , std = IMAGENET_DEFAULT_STD
489494 ),
495+ 'beitv2_large_patch16_224.in1k_ft_in1k' : _cfg (
496+ url = 'https://conversationhub.blob.core.windows.net/beit-share-public/beitv2/beitv2_large_patch16_224_pt1k_ft1k.pth' ,
497+ hf_hub_id = 'timm/' ,
498+ crop_pct = 0.95 , mean = IMAGENET_DEFAULT_MEAN , std = IMAGENET_DEFAULT_STD
499+ ),
490500 'beitv2_large_patch16_224.in1k_ft_in22k' : _cfg (
491501 url = 'https://conversationhub.blob.core.windows.net/beit-share-public/beitv2/beitv2_large_patch16_224_pt1k_ft21k.pth' ,
492502 hf_hub_id = 'timm/' ,
@@ -515,7 +525,7 @@ def _create_beit(variant, pretrained=False, **kwargs):
515525
516526
517527@register_model
518- def beit_base_patch16_224 (pretrained = False , ** kwargs ):
528+ def beit_base_patch16_224 (pretrained = False , ** kwargs ) -> Beit :
519529 model_args = dict (
520530 patch_size = 16 , embed_dim = 768 , depth = 12 , num_heads = 12 , mlp_ratio = 4 ,
521531 use_abs_pos_emb = False , use_rel_pos_bias = True , init_values = 0.1 )
@@ -524,7 +534,7 @@ def beit_base_patch16_224(pretrained=False, **kwargs):
524534
525535
526536@register_model
527- def beit_base_patch16_384 (pretrained = False , ** kwargs ):
537+ def beit_base_patch16_384 (pretrained = False , ** kwargs ) -> Beit :
528538 model_args = dict (
529539 img_size = 384 , patch_size = 16 , embed_dim = 768 , depth = 12 , num_heads = 12 ,
530540 use_abs_pos_emb = False , use_rel_pos_bias = True , init_values = 0.1 )
@@ -533,7 +543,7 @@ def beit_base_patch16_384(pretrained=False, **kwargs):
533543
534544
535545@register_model
536- def beit_large_patch16_224 (pretrained = False , ** kwargs ):
546+ def beit_large_patch16_224 (pretrained = False , ** kwargs ) -> Beit :
537547 model_args = dict (
538548 patch_size = 16 , embed_dim = 1024 , depth = 24 , num_heads = 16 ,
539549 use_abs_pos_emb = False , use_rel_pos_bias = True , init_values = 1e-5 )
@@ -542,7 +552,7 @@ def beit_large_patch16_224(pretrained=False, **kwargs):
542552
543553
544554@register_model
545- def beit_large_patch16_384 (pretrained = False , ** kwargs ):
555+ def beit_large_patch16_384 (pretrained = False , ** kwargs ) -> Beit :
546556 model_args = dict (
547557 img_size = 384 , patch_size = 16 , embed_dim = 1024 , depth = 24 , num_heads = 16 ,
548558 use_abs_pos_emb = False , use_rel_pos_bias = True , init_values = 1e-5 )
@@ -551,7 +561,7 @@ def beit_large_patch16_384(pretrained=False, **kwargs):
551561
552562
553563@register_model
554- def beit_large_patch16_512 (pretrained = False , ** kwargs ):
564+ def beit_large_patch16_512 (pretrained = False , ** kwargs ) -> Beit :
555565 model_args = dict (
556566 img_size = 512 , patch_size = 16 , embed_dim = 1024 , depth = 24 , num_heads = 16 ,
557567 use_abs_pos_emb = False , use_rel_pos_bias = True , init_values = 1e-5 )
@@ -560,7 +570,7 @@ def beit_large_patch16_512(pretrained=False, **kwargs):
560570
561571
562572@register_model
563- def beitv2_base_patch16_224 (pretrained = False , ** kwargs ):
573+ def beitv2_base_patch16_224 (pretrained = False , ** kwargs ) -> Beit :
564574 model_args = dict (
565575 patch_size = 16 , embed_dim = 768 , depth = 12 , num_heads = 12 , mlp_ratio = 4 ,
566576 use_abs_pos_emb = False , use_rel_pos_bias = True , init_values = 1e-5 )
@@ -569,7 +579,7 @@ def beitv2_base_patch16_224(pretrained=False, **kwargs):
569579
570580
571581@register_model
572- def beitv2_large_patch16_224 (pretrained = False , ** kwargs ):
582+ def beitv2_large_patch16_224 (pretrained = False , ** kwargs ) -> Beit :
573583 model_args = dict (
574584 patch_size = 16 , embed_dim = 1024 , depth = 24 , num_heads = 16 ,
575585 use_abs_pos_emb = False , use_rel_pos_bias = True , init_values = 1e-5 )
0 commit comments