@@ -93,7 +93,9 @@ def _cfg(url='', **kwargs):
9393 ),
9494
9595 gmixer_12_224 = _cfg (mean = IMAGENET_DEFAULT_MEAN , std = IMAGENET_DEFAULT_STD ),
96- gmixer_24_224 = _cfg (mean = IMAGENET_DEFAULT_MEAN , std = IMAGENET_DEFAULT_STD ),
96+ gmixer_24_224 = _cfg (
97+ url = 'https://github.com/rwightman/pytorch-image-models/releases/download/v0.1-weights/gmixer_24_224_raa-7daf7ae6.pth' ,
98+ mean = IMAGENET_DEFAULT_MEAN , std = IMAGENET_DEFAULT_STD ),
9799
98100 resmlp_12_224 = _cfg (
99101 url = 'https://dl.fbaipublicfiles.com/deit/resmlp_12_no_dist.pth' ,
@@ -457,23 +459,23 @@ def mixer_b16_224_miil_in21k(pretrained=False, **kwargs):
457459
458460@register_model
459461def gmixer_12_224 (pretrained = False , ** kwargs ):
460- """ Glu-Mixer-12 224x224 (short & fat)
462+ """ Glu-Mixer-12 224x224
461463 Experiment by Ross Wightman, adding (Si)GLU to MLP-Mixer
462464 """
463465 model_args = dict (
464- patch_size = 16 , num_blocks = 12 , embed_dim = 512 , mlp_ratio = (1.0 , 6 .0 ),
466+ patch_size = 16 , num_blocks = 12 , embed_dim = 384 , mlp_ratio = (1.0 , 4 .0 ),
465467 mlp_layer = GluMlp , act_layer = nn .SiLU , ** kwargs )
466468 model = _create_mixer ('gmixer_12_224' , pretrained = pretrained , ** model_args )
467469 return model
468470
469471
470472@register_model
471473def gmixer_24_224 (pretrained = False , ** kwargs ):
472- """ Glu-Mixer-24 224x224 (tall & slim)
474+ """ Glu-Mixer-24 224x224
473475 Experiment by Ross Wightman, adding (Si)GLU to MLP-Mixer
474476 """
475477 model_args = dict (
476- patch_size = 16 , num_blocks = 24 , embed_dim = 384 , mlp_ratio = (1.0 , 6 .0 ),
478+ patch_size = 16 , num_blocks = 24 , embed_dim = 384 , mlp_ratio = (1.0 , 4 .0 ),
477479 mlp_layer = GluMlp , act_layer = nn .SiLU , ** kwargs )
478480 model = _create_mixer ('gmixer_24_224' , pretrained = pretrained , ** model_args )
479481 return model
0 commit comments