@@ -2282,107 +2282,91 @@ def _cfgr(url='', **kwargs):
22822282 # original attention pool head variants
22832283 'resnet50_clip.openai' : _cfgr (
22842284 hf_hub_id = 'timm/' ,
2285- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
22862285 num_classes = 1024 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
22872286 fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
22882287 classifier = 'head.proj' ,
22892288 ),
22902289 'resnet101_clip.openai' : _cfgr (
22912290 hf_hub_id = 'timm/' ,
2292- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
22932291 num_classes = 512 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
22942292 fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
22952293 classifier = 'head.proj' ,
22962294 ),
22972295 'resnet50x4_clip.openai' : _cfgr (
22982296 hf_hub_id = 'timm/' ,
2299- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
23002297 num_classes = 640 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23012298 fixed_input_size = True , input_size = (3 , 288 , 288 ), pool_size = (9 , 9 ),
23022299 classifier = 'head.proj' ,
23032300 ),
23042301 'resnet50x16_clip.openai' : _cfgr (
23052302 hf_hub_id = 'timm/' ,
2306- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
23072303 num_classes = 768 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23082304 fixed_input_size = True , input_size = (3 , 384 , 384 ), pool_size = (12 , 12 ),
23092305 classifier = 'head.proj' ,
23102306 ),
23112307 'resnet50x64_clip.openai' : _cfgr (
23122308 hf_hub_id = 'timm/' ,
2313- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
23142309 num_classes = 1024 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23152310 fixed_input_size = True , input_size = (3 , 448 , 448 ), pool_size = (14 , 14 ),
23162311 classifier = 'head.proj' ,
23172312 ),
23182313 'resnet50_clip.cc12m' : _cfgr (
23192314 hf_hub_id = 'timm/' ,
2320- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
23212315 num_classes = 1024 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23222316 fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
23232317 classifier = 'head.proj' ,
23242318 ),
23252319 'resnet50_clip.yfcc15m' : _cfgr (
23262320 hf_hub_id = 'timm/' ,
2327- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
23282321 num_classes = 1024 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23292322 fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
23302323 classifier = 'head.proj' ,
23312324 ),
23322325 'resnet101_clip.yfcc15m' : _cfgr (
23332326 hf_hub_id = 'timm/' ,
2334- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
23352327 num_classes = 512 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23362328 fixed_input_size = True , input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
23372329 classifier = 'head.proj' ,
23382330 ),
23392331
23402332 # avg-pool w/ optional standard classifier head variants
23412333 'resnet50_clip_gap.openai' : _cfgr (
2342- hf_hub_id = 'timm/resnet50_clip.openai' ,
2343- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2334+ hf_hub_id = 'timm/' ,
23442335 num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23452336 input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
23462337 ),
23472338 'resnet101_clip_gap.openai' : _cfgr (
2348- hf_hub_id = 'timm/resnet101_clip.openai' ,
2349- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2339+ hf_hub_id = 'timm/' ,
23502340 num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23512341 input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
23522342 ),
23532343 'resnet50x4_clip_gap.openai' : _cfgr (
2354- hf_hub_id = 'timm/resnet50x4_clip.openai' ,
2355- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2344+ hf_hub_id = 'timm/' ,
23562345 num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23572346 input_size = (3 , 288 , 288 ), pool_size = (9 , 9 ),
23582347 ),
23592348 'resnet50x16_clip_gap.openai' : _cfgr (
2360- hf_hub_id = 'timm/resnet50x16_clip.openai' ,
2361- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2349+ hf_hub_id = 'timm/' ,
23622350 num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23632351 input_size = (3 , 384 , 384 ), pool_size = (12 , 12 ),
23642352 ),
23652353 'resnet50x64_clip_gap.openai' : _cfgr (
2366- hf_hub_id = 'timm/resnet50x64_clip.openai' ,
2367- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2354+ hf_hub_id = 'timm/' ,
23682355 num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23692356 input_size = (3 , 448 , 448 ), pool_size = (14 , 14 ),
23702357 ),
23712358 'resnet50_clip_gap.cc12m' : _cfgr (
2372- hf_hub_id = 'timm/resnet50_clip.cc12m' ,
2373- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2359+ hf_hub_id = 'timm/' ,
23742360 num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23752361 input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
23762362 ),
23772363 'resnet50_clip_gap.yfcc15m' : _cfgr (
2378- hf_hub_id = 'timm/resnet50_clip.yfcc15m' ,
2379- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2364+ hf_hub_id = 'timm/' ,
23802365 num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23812366 input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
23822367 ),
23832368 'resnet101_clip_gap.yfcc15m' : _cfgr (
2384- hf_hub_id = 'timm/resnet101_clip.yfcc15m' ,
2385- hf_hub_filename = 'open_clip_pytorch_model.bin' ,
2369+ hf_hub_id = 'timm/' ,
23862370 num_classes = 0 , mean = OPENAI_CLIP_MEAN , std = OPENAI_CLIP_STD ,
23872371 input_size = (3 , 224 , 224 ), pool_size = (7 , 7 ),
23882372 ),
0 commit comments