@@ -1556,9 +1556,6 @@ def _cfg(url: str = '', **kwargs) -> Dict[str, Any]:
1556
1556
hf_hub_id = 'timm/' ,
1557
1557
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 ),
1558
1558
1559
- 'vit_base_patch32_clip_224.laion2b_ft_in12k' : _cfg (
1560
- #hf_hub_id='timm/vit_base_patch32_clip_224.laion2b_ft_in12k', # FIXME weight exists, need to push
1561
- mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , num_classes = 11821 ),
1562
1559
'vit_base_patch16_clip_224.laion2b_ft_in12k' : _cfg (
1563
1560
hf_hub_id = 'timm/' ,
1564
1561
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , num_classes = 11821 ),
@@ -1569,9 +1566,6 @@ def _cfg(url: str = '', **kwargs) -> Dict[str, Any]:
1569
1566
hf_hub_id = 'timm/' ,
1570
1567
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 11821 ),
1571
1568
1572
- 'vit_base_patch32_clip_224.openai_ft_in12k' : _cfg (
1573
- # hf_hub_id='timm/vit_base_patch32_clip_224.openai_ft_in12k', # FIXME weight exists, need to push
1574
- mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , num_classes = 11821 ),
1575
1569
'vit_base_patch16_clip_224.openai_ft_in12k' : _cfg (
1576
1570
hf_hub_id = 'timm/' ,
1577
1571
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , num_classes = 11821 ),
@@ -1580,28 +1574,22 @@ def _cfg(url: str = '', **kwargs) -> Dict[str, Any]:
1580
1574
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 11821 ),
1581
1575
1582
1576
'vit_base_patch32_clip_224.laion2b' : _cfg (
1583
- hf_hub_id = 'laion/CLIP-ViT-B-32-laion2B-s34B-b79K' ,
1584
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1577
+ hf_hub_id = 'timm/' ,
1585
1578
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , num_classes = 512 ),
1586
1579
'vit_base_patch16_clip_224.laion2b' : _cfg (
1587
- hf_hub_id = 'laion/CLIP-ViT-B-16-laion2B-s34B-b88K' ,
1588
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1580
+ hf_hub_id = 'timm/' ,
1589
1581
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 512 ),
1590
1582
'vit_large_patch14_clip_224.laion2b' : _cfg (
1591
- hf_hub_id = 'laion/CLIP-ViT-L-14-laion2B-s32B-b82K' ,
1592
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1583
+ hf_hub_id = 'timm/' ,
1593
1584
mean = IMAGENET_INCEPTION_MEAN , std = IMAGENET_INCEPTION_STD , crop_pct = 1.0 , num_classes = 768 ),
1594
1585
'vit_huge_patch14_clip_224.laion2b' : _cfg (
1595
- hf_hub_id = 'laion/CLIP-ViT-H-14-laion2B-s32B-b79K' ,
1596
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1586
+ hf_hub_id = 'timm/' ,
1597
1587
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 1024 ),
1598
1588
'vit_giant_patch14_clip_224.laion2b' : _cfg (
1599
- hf_hub_id = 'laion/CLIP-ViT-g-14-laion2B-s12B-b42K' ,
1600
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1589
+ hf_hub_id = 'timm/' ,
1601
1590
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 1024 ),
1602
1591
'vit_gigantic_patch14_clip_224.laion2b' : _cfg (
1603
- hf_hub_id = 'laion/CLIP-ViT-bigG-14-laion2B-39B-b160k' ,
1604
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1592
+ hf_hub_id = 'timm/' ,
1605
1593
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 1280 ),
1606
1594
1607
1595
'vit_base_patch32_clip_224.laion400m_e32' : _cfg (
@@ -1620,21 +1608,17 @@ def _cfg(url: str = '', **kwargs) -> Dict[str, Any]:
1620
1608
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 768 ),
1621
1609
1622
1610
'vit_base_patch32_clip_224.datacompxl' : _cfg (
1623
- hf_hub_id = 'laion/CLIP-ViT-B-32-DataComp.XL-s13B-b90K' ,
1624
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1611
+ hf_hub_id = 'timm/' ,
1625
1612
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 512 ),
1626
1613
'vit_base_patch32_clip_256.datacompxl' : _cfg (
1627
- hf_hub_id = 'laion/CLIP-ViT-B-32-256x256-DataComp-s34B-b86K' ,
1628
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1614
+ hf_hub_id = 'timm/' ,
1629
1615
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
1630
1616
crop_pct = 1.0 , input_size = (3 , 256 , 256 ), num_classes = 512 ),
1631
1617
'vit_base_patch16_clip_224.datacompxl' : _cfg (
1632
- hf_hub_id = 'laion/CLIP-ViT-B-16-DataComp.XL-s13B-b90K' ,
1633
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1618
+ hf_hub_id = 'timm/' ,
1634
1619
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 512 ),
1635
1620
'vit_large_patch14_clip_224.datacompxl' : _cfg (
1636
- hf_hub_id = 'laion/CLIP-ViT-L-14-DataComp.XL-s13B-b90K' ,
1637
- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1621
+ hf_hub_id = 'timm/' ,
1638
1622
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 768 ),
1639
1623
1640
1624
'vit_base_patch16_clip_224.dfn2b' : _cfg (
@@ -1659,42 +1643,46 @@ def _cfg(url: str = '', **kwargs) -> Dict[str, Any]:
1659
1643
crop_pct = 1.0 , input_size = (3 , 378 , 378 ), num_classes = 1024 ),
1660
1644
1661
1645
'vit_base_patch32_clip_224.metaclip_2pt5b' : _cfg (
1662
- hf_hub_id = 'timm/' , hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1646
+ hf_hub_id = 'timm/' ,
1663
1647
license = 'cc-by-nc-4.0' ,
1664
1648
notes = ('natively QuickGELU, use quickgelu model variant for original results' ,),
1665
1649
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 512 ),
1666
1650
'vit_base_patch16_clip_224.metaclip_2pt5b' : _cfg (
1667
- hf_hub_id = 'timm/' , hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1651
+ hf_hub_id = 'timm/' ,
1668
1652
license = 'cc-by-nc-4.0' ,
1669
1653
notes = ('natively QuickGELU, use quickgelu model variant for original results' ,),
1670
1654
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 512 ),
1671
1655
'vit_large_patch14_clip_224.metaclip_2pt5b' : _cfg (
1672
- hf_hub_id = 'timm/' , hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1656
+ hf_hub_id = 'timm/' ,
1673
1657
license = 'cc-by-nc-4.0' ,
1674
1658
notes = ('natively QuickGELU, use quickgelu model variant for original results' ,),
1675
1659
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 768 ),
1676
1660
'vit_huge_patch14_clip_224.metaclip_2pt5b' : _cfg (
1677
- hf_hub_id = 'timm/' , hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1661
+ hf_hub_id = 'timm/' ,
1678
1662
license = 'cc-by-nc-4.0' ,
1679
1663
notes = ('natively QuickGELU, use quickgelu model variant for original results' ,),
1680
1664
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 1024 ),
1665
+ 'vit_huge_patch14_clip_224.metaclip_altogether' : _cfg (
1666
+ hf_hub_id = 'timm/' ,
1667
+ license = 'cc-by-nc-4.0' ,
1668
+ mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 1024 ),
1681
1669
'vit_gigantic_patch14_clip_224.metaclip_2pt5b' : _cfg (
1682
- hf_hub_id = 'timm/' , hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1670
+ hf_hub_id = 'timm/' ,
1683
1671
license = 'cc-by-nc-4.0' ,
1684
1672
notes = ('natively QuickGELU, use quickgelu model variant for original results' ,),
1685
1673
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 1280 ),
1686
1674
'vit_base_patch32_clip_224.metaclip_400m' : _cfg (
1687
- hf_hub_id = 'timm/' , hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1675
+ hf_hub_id = 'timm/' ,
1688
1676
license = 'cc-by-nc-4.0' ,
1689
1677
notes = ('natively QuickGELU, use quickgelu model variant for original results' ,),
1690
1678
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 512 ),
1691
1679
'vit_base_patch16_clip_224.metaclip_400m' : _cfg (
1692
- hf_hub_id = 'timm/' , hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1680
+ hf_hub_id = 'timm/' ,
1693
1681
license = 'cc-by-nc-4.0' ,
1694
1682
notes = ('natively QuickGELU, use quickgelu model variant for original results' ,),
1695
1683
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 512 ),
1696
1684
'vit_large_patch14_clip_224.metaclip_400m' : _cfg (
1697
- hf_hub_id = 'timm/' , hf_hub_filename = 'open_clip_pytorch_model.bin' ,
1685
+ hf_hub_id = 'timm/' ,
1698
1686
license = 'cc-by-nc-4.0' ,
1699
1687
notes = ('natively QuickGELU, use quickgelu model variant for original results' ,),
1700
1688
mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD , crop_pct = 1.0 , num_classes = 768 ),
0 commit comments