brainscore-vision 2.2.3__py3-none-any.whl → 2.2.5__py3-none-any.whl
Sign up to get free protection for your applications and to get access to all the features.
- brainscore_vision/data/baker2022/__init__.py +10 -10
- brainscore_vision/data/baker2022/data_packaging/inverted_distortion_data_assembly.py +2 -2
- brainscore_vision/data/baker2022/data_packaging/inverted_distortion_stimulus_set.py +2 -2
- brainscore_vision/data/baker2022/data_packaging/normal_distortion_data_assembly.py +2 -2
- brainscore_vision/data/baker2022/data_packaging/normal_distortion_stimulus_set.py +2 -2
- brainscore_vision/data/barbumayo2019/__init__.py +3 -3
- brainscore_vision/data/bashivankar2019/__init__.py +10 -10
- brainscore_vision/data/bashivankar2019/data_packaging/synthetic.py +2 -2
- brainscore_vision/data/bmd2024/__init__.py +20 -20
- brainscore_vision/data/bmd2024/data_packaging/BMD_2024_data_assembly.py +2 -1
- brainscore_vision/data/bmd2024/data_packaging/BMD_2024_simulus_set.py +2 -1
- brainscore_vision/data/bracci2019/__init__.py +5 -5
- brainscore_vision/data/bracci2019/data_packaging.py +1 -1
- brainscore_vision/data/cadena2017/__init__.py +5 -5
- brainscore_vision/data/cichy2019/__init__.py +5 -5
- brainscore_vision/data/coggan2024_behavior/__init__.py +8 -8
- brainscore_vision/data/coggan2024_behavior/data_packaging.py +2 -2
- brainscore_vision/data/coggan2024_fMRI/__init__.py +5 -6
- brainscore_vision/data/coggan2024_fMRI/data_packaging.py +2 -2
- brainscore_vision/data/david2004/__init__.py +5 -5
- brainscore_vision/data/deng2009/__init__.py +3 -3
- brainscore_vision/data/ferguson2024/__init__.py +112 -112
- brainscore_vision/data/ferguson2024/data_packaging/data_packaging.py +2 -2
- brainscore_vision/data/freemanziemba2013/__init__.py +31 -30
- brainscore_vision/data/geirhos2021/__init__.py +85 -85
- brainscore_vision/data/geirhos2021/data_packaging/colour/colour_data_assembly.py +2 -2
- brainscore_vision/data/geirhos2021/data_packaging/colour/colour_stimulus_set.py +2 -2
- brainscore_vision/data/geirhos2021/data_packaging/contrast/contrast_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/contrast/contrast_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/cue-conflict/cue-conflict_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/cue-conflict/cue-conflict_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/edge/edge_data_assembly.py +2 -2
- brainscore_vision/data/geirhos2021/data_packaging/edge/edge_stimulus_set.py +2 -2
- brainscore_vision/data/geirhos2021/data_packaging/eidolonI/eidolonI_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/eidolonI/eidolonI_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/eidolonII/eidolonII_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/eidolonII/eidolonII_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/eidolonIII/eidolonIII_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/eidolonIII/eidolonIII_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/false-colour/false-colour_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/false-colour/false-colour_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/high-pass/high-pass_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/high-pass/high-pass_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/low-pass/low-pass_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/low-pass/low-pass_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/phase-scrambling/phase-scrambling_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/phase-scrambling/phase-scrambling_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/power-equalisation/power-equalisation_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/power-equalisation/power-equalisation_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/rotation/rotation_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/rotation/rotation_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/silhouette/silhouette_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/silhouette/silhouette_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/sketch/sketch_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/sketch/sketch_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/stylized/stylized_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/stylized/stylized_stimulus_set.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/uniform-noise/uniform-noise_data_assembly.py +1 -1
- brainscore_vision/data/geirhos2021/data_packaging/uniform-noise/uniform-noise_stimulus_set.py +1 -1
- brainscore_vision/data/hebart2023/__init__.py +5 -5
- brainscore_vision/data/hebart2023/packaging/data_assembly.py +2 -2
- brainscore_vision/data/hebart2023/packaging/stimulus_set.py +1 -1
- brainscore_vision/data/hendrycks2019/__init__.py +12 -12
- brainscore_vision/data/igustibagus2024/__init__.py +5 -5
- brainscore_vision/data/imagenetslim15000/__init__.py +3 -3
- brainscore_vision/data/islam2021/__init__.py +3 -3
- brainscore_vision/data/kar2018/__init__.py +7 -7
- brainscore_vision/data/kar2019/__init__.py +5 -5
- brainscore_vision/data/kuzovkin2018/__init__.py +5 -5
- brainscore_vision/data/lonnqvist2024/__init__.py +12 -12
- brainscore_vision/data/lonnqvist2024/data_packaging/lonnqvist_data_assembly.py +1 -1
- brainscore_vision/data/lonnqvist2024/data_packaging/lonnqvist_stimulus_set.py +1 -1
- brainscore_vision/data/majajhong2015/__init__.py +23 -23
- brainscore_vision/data/malania2007/__init__.py +77 -77
- brainscore_vision/data/malania2007/data_packaging/malania_data_assembly.py +1 -1
- brainscore_vision/data/malania2007/data_packaging/malania_stimulus_set.py +1 -1
- brainscore_vision/data/maniquet2024/__init__.py +11 -11
- brainscore_vision/data/marques2020/__init__.py +30 -30
- brainscore_vision/data/rajalingham2018/__init__.py +10 -10
- brainscore_vision/data/rajalingham2020/__init__.py +5 -5
- brainscore_vision/data/rust2012/__init__.py +7 -7
- brainscore_vision/data/sanghavi2020/__init__.py +19 -19
- brainscore_vision/data/scialom2024/__init__.py +110 -110
- brainscore_vision/data/scialom2024/data_packaging/scialom_data_assembly.py +1 -1
- brainscore_vision/data/scialom2024/data_packaging/scialom_stimulus_set.py +1 -1
- brainscore_vision/data/seibert2019/__init__.py +2 -2
- brainscore_vision/data/zhang2018/__init__.py +5 -5
- brainscore_vision/data_helpers/s3.py +25 -6
- brainscore_vision/model_helpers/activations/pytorch.py +34 -12
- brainscore_vision/models/AT_efficientnet_b2/__init__.py +7 -0
- brainscore_vision/models/AT_efficientnet_b2/model.py +58 -0
- brainscore_vision/models/AT_efficientnet_b2/region_layer_map/AT_efficientnet-b2.json +6 -0
- brainscore_vision/models/AT_efficientnet_b2/requirements.txt +1 -0
- brainscore_vision/models/AT_efficientnet_b2/test.py +8 -0
- brainscore_vision/models/AdvProp_efficientnet_b2/__init__.py +7 -0
- brainscore_vision/models/AdvProp_efficientnet_b2/model.py +64 -0
- brainscore_vision/models/AdvProp_efficientnet_b2/region_layer_map/AdvProp_efficientnet-b2.json +1 -0
- brainscore_vision/models/AdvProp_efficientnet_b2/requirements.txt +1 -0
- brainscore_vision/models/AdvProp_efficientnet_b2/test.py +8 -0
- brainscore_vision/models/AdvProp_efficientnet_b4/__init__.py +5 -0
- brainscore_vision/models/AdvProp_efficientnet_b4/model.py +65 -0
- brainscore_vision/models/AdvProp_efficientnet_b4/region_layer_map/AdvProp_efficientnet-b4.json +1 -0
- brainscore_vision/models/AdvProp_efficientnet_b4/requirements.txt +1 -0
- brainscore_vision/models/AdvProp_efficientnet_b4/test.py +8 -0
- brainscore_vision/models/AdvProp_efficientnet_b7/__init__.py +5 -0
- brainscore_vision/models/AdvProp_efficientnet_b7/model.py +65 -0
- brainscore_vision/models/AdvProp_efficientnet_b7/region_layer_map/AdvProp_efficientnet-b7.json +1 -0
- brainscore_vision/models/AdvProp_efficientnet_b7/requirements.txt +1 -0
- brainscore_vision/models/AdvProp_efficientnet_b7/test.py +8 -0
- brainscore_vision/models/AdvProp_efficientnet_b8/__init__.py +7 -0
- brainscore_vision/models/AdvProp_efficientnet_b8/model.py +65 -0
- brainscore_vision/models/AdvProp_efficientnet_b8/region_layer_map/AdvProp_efficientnet-b8.json +1 -0
- brainscore_vision/models/AdvProp_efficientnet_b8/requirements.txt +1 -0
- brainscore_vision/models/AdvProp_efficientnet_b8/test.py +8 -0
- brainscore_vision/models/BiT_S_R101x1/__init__.py +7 -0
- brainscore_vision/models/BiT_S_R101x1/model.py +223 -0
- brainscore_vision/models/BiT_S_R101x1/region_layer_map/BiT-S-R101x1.json +1 -0
- brainscore_vision/models/BiT_S_R101x1/requirements.txt +4 -0
- brainscore_vision/models/BiT_S_R101x1/test.py +8 -0
- brainscore_vision/models/BiT_S_R101x3/__init__.py +7 -0
- brainscore_vision/models/BiT_S_R101x3/model.py +225 -0
- brainscore_vision/models/BiT_S_R101x3/region_layer_map/BiT-S-R101x3.json +1 -0
- brainscore_vision/models/BiT_S_R101x3/requirements.txt +4 -0
- brainscore_vision/models/BiT_S_R101x3/test.py +8 -0
- brainscore_vision/models/BiT_S_R152x2/__init__.py +7 -0
- brainscore_vision/models/BiT_S_R152x2/model.py +231 -0
- brainscore_vision/models/BiT_S_R152x2/region_layer_map/BiT-S-R152x2.json +1 -0
- brainscore_vision/models/BiT_S_R152x2/requirements.txt +4 -0
- brainscore_vision/models/BiT_S_R152x2/test.py +8 -0
- brainscore_vision/models/BiT_S_R152x4/__init__.py +7 -0
- brainscore_vision/models/BiT_S_R152x4/model.py +231 -0
- brainscore_vision/models/BiT_S_R152x4/region_layer_map/BiT-S-R152x4.json +1 -0
- brainscore_vision/models/BiT_S_R152x4/requirements.txt +4 -0
- brainscore_vision/models/BiT_S_R152x4/test.py +8 -0
- brainscore_vision/models/BiT_S_R50x1/__init__.py +7 -0
- brainscore_vision/models/BiT_S_R50x1/model.py +218 -0
- brainscore_vision/models/BiT_S_R50x1/region_layer_map/BiT-S-R50x1.json +1 -0
- brainscore_vision/models/BiT_S_R50x1/requirements.txt +4 -0
- brainscore_vision/models/BiT_S_R50x1/test.py +8 -0
- brainscore_vision/models/BiT_S_R50x3/__init__.py +7 -0
- brainscore_vision/models/BiT_S_R50x3/model.py +217 -0
- brainscore_vision/models/BiT_S_R50x3/region_layer_map/BiT-S-R50x3.json +1 -0
- brainscore_vision/models/BiT_S_R50x3/requirements.txt +4 -0
- brainscore_vision/models/BiT_S_R50x3/test.py +8 -0
- brainscore_vision/models/ReAlnet/__init__.py +64 -0
- brainscore_vision/models/ReAlnet/model.py +237 -0
- brainscore_vision/models/ReAlnet/requirements.txt +7 -0
- brainscore_vision/models/ReAlnet/test.py +0 -0
- brainscore_vision/models/ReAlnet/weights.json +26 -0
- brainscore_vision/models/ReAlnet_cornet/__init__.py +46 -0
- brainscore_vision/models/ReAlnet_cornet/helpers/helpers.py +215 -0
- brainscore_vision/models/ReAlnet_cornet/model.py +69 -0
- brainscore_vision/models/ReAlnet_cornet/requirements.txt +8 -0
- brainscore_vision/models/ReAlnet_cornet/test.py +0 -0
- brainscore_vision/models/Res2Net50_26w_4s/__init__.py +5 -0
- brainscore_vision/models/Res2Net50_26w_4s/helpers/resnet_helpers.py +161 -0
- brainscore_vision/models/Res2Net50_26w_4s/model.py +75 -0
- brainscore_vision/models/Res2Net50_26w_4s/region_layer_map/Res2Net50_26w_4s.json +1 -0
- brainscore_vision/models/Res2Net50_26w_4s/requirements.txt +1 -0
- brainscore_vision/models/Res2Net50_26w_4s/test.py +8 -0
- brainscore_vision/models/VOneCORnet_S/__init__.py +9 -0
- brainscore_vision/models/VOneCORnet_S/helpers/cornet_helpers.py +34 -0
- brainscore_vision/models/VOneCORnet_S/helpers/cornet_s_helpers.py +128 -0
- brainscore_vision/models/VOneCORnet_S/helpers/cornets.py +136 -0
- brainscore_vision/models/VOneCORnet_S/helpers/vonecornets.py +38 -0
- brainscore_vision/models/VOneCORnet_S/model.py +25 -0
- brainscore_vision/models/VOneCORnet_S/requirements.txt +1 -0
- brainscore_vision/models/VOneCORnet_S/test.py +8 -0
- brainscore_vision/models/alexnet_training_seed_01/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_01/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_01/region_layer_map/alexnet_training_seed_01.json +6 -0
- brainscore_vision/models/alexnet_training_seed_01/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_01/test.py +9 -0
- brainscore_vision/models/alexnet_training_seed_02/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_02/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_02/region_layer_map/alexnet_training_seed_02.json +6 -0
- brainscore_vision/models/alexnet_training_seed_02/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_02/test.py +9 -0
- brainscore_vision/models/alexnet_training_seed_03/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_03/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_03/region_layer_map/alexnet_training_seed_03.json +6 -0
- brainscore_vision/models/alexnet_training_seed_03/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_03/test.py +9 -0
- brainscore_vision/models/alexnet_training_seed_04/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_04/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_04/region_layer_map/alexnet_training_seed_04.json +6 -0
- brainscore_vision/models/alexnet_training_seed_04/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_04/test.py +9 -0
- brainscore_vision/models/alexnet_training_seed_05/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_05/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_05/region_layer_map/alexnet_training_seed_05.json +6 -0
- brainscore_vision/models/alexnet_training_seed_05/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_05/test.py +9 -0
- brainscore_vision/models/alexnet_training_seed_06/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_06/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_06/region_layer_map/alexnet_training_seed_06.json +6 -0
- brainscore_vision/models/alexnet_training_seed_06/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_06/test.py +9 -0
- brainscore_vision/models/alexnet_training_seed_07/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_07/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_07/region_layer_map/alexnet_training_seed_07.json +6 -0
- brainscore_vision/models/alexnet_training_seed_07/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_07/test.py +9 -0
- brainscore_vision/models/alexnet_training_seed_08/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_08/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_08/region_layer_map/alexnet_training_seed_08.json +6 -0
- brainscore_vision/models/alexnet_training_seed_08/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_08/test.py +9 -0
- brainscore_vision/models/alexnet_training_seed_09/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_09/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_09/region_layer_map/alexnet_training_seed_09.json +6 -0
- brainscore_vision/models/alexnet_training_seed_09/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_09/test.py +9 -0
- brainscore_vision/models/alexnet_training_seed_10/__init__.py +6 -0
- brainscore_vision/models/alexnet_training_seed_10/model.py +140 -0
- brainscore_vision/models/alexnet_training_seed_10/region_layer_map/alexnet_training_seed_10.json +6 -0
- brainscore_vision/models/alexnet_training_seed_10/requirements.txt +3 -0
- brainscore_vision/models/alexnet_training_seed_10/test.py +9 -0
- brainscore_vision/models/antialiased-r50/__init__.py +7 -0
- brainscore_vision/models/antialiased-r50/model.py +62 -0
- brainscore_vision/models/antialiased-r50/region_layer_map/antialiased-r50.json +1 -0
- brainscore_vision/models/antialiased-r50/requirements.txt +3 -0
- brainscore_vision/models/antialiased-r50/test.py +8 -0
- brainscore_vision/models/convnext_tiny_sup/__init__.py +8 -0
- brainscore_vision/models/convnext_tiny_sup/model.py +56 -0
- brainscore_vision/models/convnext_tiny_sup/region_layer_map/convnext_tiny_sup.json +1 -0
- brainscore_vision/models/convnext_tiny_sup/requirements.txt +1 -0
- brainscore_vision/models/convnext_tiny_sup/test.py +8 -0
- brainscore_vision/models/cornet_s/model.py +2 -2
- brainscore_vision/models/custom_model_cv_18_dagger_408/model.py +2 -2
- brainscore_vision/models/densenet_121/__init__.py +7 -0
- brainscore_vision/models/densenet_121/model.py +63 -0
- brainscore_vision/models/densenet_121/region_layer_map/densenet-121.json +1 -0
- brainscore_vision/models/densenet_121/requirements.txt +1 -0
- brainscore_vision/models/densenet_121/test.py +8 -0
- brainscore_vision/models/densenet_169/__init__.py +7 -0
- brainscore_vision/models/densenet_169/model.py +63 -0
- brainscore_vision/models/densenet_169/region_layer_map/densenet-169.json +1 -0
- brainscore_vision/models/densenet_169/requirements.txt +1 -0
- brainscore_vision/models/densenet_169/test.py +9 -0
- brainscore_vision/models/{densenet_201_pytorch → densenet_201}/__init__.py +3 -3
- brainscore_vision/models/{densenet_201_pytorch → densenet_201}/model.py +12 -10
- brainscore_vision/models/densenet_201/region_layer_map/densenet-201.json +6 -0
- brainscore_vision/models/densenet_201/test.py +8 -0
- brainscore_vision/models/efficientnet_b0/__init__.py +7 -0
- brainscore_vision/models/efficientnet_b0/model.py +45 -0
- brainscore_vision/models/efficientnet_b0/region_layer_map/efficientnet_b0.json +1 -0
- brainscore_vision/models/efficientnet_b0/requirements.txt +2 -0
- brainscore_vision/models/efficientnet_b0/test.py +8 -0
- brainscore_vision/models/efficientnet_b7/__init__.py +7 -0
- brainscore_vision/models/efficientnet_b7/model.py +61 -0
- brainscore_vision/models/efficientnet_b7/region_layer_map/efficientnet-b7.json +1 -0
- brainscore_vision/models/efficientnet_b7/requirements.txt +1 -0
- brainscore_vision/models/efficientnet_b7/test.py +9 -0
- brainscore_vision/models/effnetb1_cutmix_augmix_sam_e1_5avg_424x377/model.py +2 -2
- brainscore_vision/models/effnetb1_cutmixpatch_SAM_robust32_avge6e8e9e10_manylayers_324x288/model.py +142 -142
- brainscore_vision/models/effnetb1_cutmixpatch_augmix_robust32_avge4e7_manylayers_324x288/model.py +2 -2
- brainscore_vision/models/evresnet_50_1/__init__.py +12 -0
- brainscore_vision/models/evresnet_50_1/evnet/backends.py +109 -0
- brainscore_vision/models/evresnet_50_1/evnet/evnet.py +147 -0
- brainscore_vision/models/evresnet_50_1/evnet/modules.py +308 -0
- brainscore_vision/models/evresnet_50_1/evnet/params.py +326 -0
- brainscore_vision/models/evresnet_50_1/evnet/utils.py +142 -0
- brainscore_vision/models/evresnet_50_1/model.py +62 -0
- brainscore_vision/models/evresnet_50_1/requirements.txt +5 -0
- brainscore_vision/models/evresnet_50_1/test.py +8 -0
- brainscore_vision/models/evresnet_50_4/__init__.py +12 -0
- brainscore_vision/models/evresnet_50_4/evnet/backends.py +109 -0
- brainscore_vision/models/evresnet_50_4/evnet/evnet.py +147 -0
- brainscore_vision/models/evresnet_50_4/evnet/modules.py +308 -0
- brainscore_vision/models/evresnet_50_4/evnet/params.py +326 -0
- brainscore_vision/models/evresnet_50_4/evnet/utils.py +142 -0
- brainscore_vision/models/evresnet_50_4/model.py +67 -0
- brainscore_vision/models/evresnet_50_4/requirements.txt +4 -0
- brainscore_vision/models/evresnet_50_4/test.py +8 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/__init__.py +10 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/evnet/backends.py +109 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/evnet/evnet.py +147 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/evnet/modules.py +308 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/evnet/params.py +326 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/evnet/utils.py +142 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/model.py +67 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/region_layer_map/evresnet_50_4_no_mapping.json +6 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/requirements.txt +4 -0
- brainscore_vision/models/evresnet_50_4_no_mapping/test.py +8 -0
- brainscore_vision/models/grcnn/__init__.py +7 -0
- brainscore_vision/models/grcnn/helpers/helpers.py +236 -0
- brainscore_vision/models/grcnn/model.py +54 -0
- brainscore_vision/models/grcnn/region_layer_map/grcnn.json +1 -0
- brainscore_vision/models/grcnn/requirements.txt +2 -0
- brainscore_vision/models/grcnn/test.py +9 -0
- brainscore_vision/models/grcnn_109/__init__.py +5 -0
- brainscore_vision/models/grcnn_109/helpers/helpers.py +237 -0
- brainscore_vision/models/grcnn_109/model.py +53 -0
- brainscore_vision/models/grcnn_109/region_layer_map/grcnn_109.json +1 -0
- brainscore_vision/models/grcnn_109/requirements.txt +2 -0
- brainscore_vision/models/grcnn_109/test.py +9 -0
- brainscore_vision/models/hmax/model.py +2 -2
- brainscore_vision/models/imagenet_l2_3_0/__init__.py +9 -0
- brainscore_vision/models/imagenet_l2_3_0/model.py +101 -0
- brainscore_vision/models/imagenet_l2_3_0/region_layer_map/imagenet_l2_3_0.json +1 -0
- brainscore_vision/models/imagenet_l2_3_0/requirements.txt +2 -0
- brainscore_vision/models/imagenet_l2_3_0/test.py +8 -0
- brainscore_vision/models/inception_v1/__init__.py +7 -0
- brainscore_vision/models/inception_v1/model.py +67 -0
- brainscore_vision/models/inception_v1/requirements.txt +1 -0
- brainscore_vision/models/inception_v1/test.py +8 -0
- brainscore_vision/models/{inception_v3_pytorch → inception_v3}/__init__.py +3 -3
- brainscore_vision/models/{inception_v3_pytorch → inception_v3}/model.py +10 -10
- brainscore_vision/models/inception_v3/region_layer_map/inception_v3.json +6 -0
- brainscore_vision/models/inception_v3/test.py +8 -0
- brainscore_vision/models/{inception_v4_pytorch → inception_v4}/__init__.py +3 -3
- brainscore_vision/models/{inception_v4_pytorch → inception_v4}/model.py +8 -15
- brainscore_vision/models/inception_v4/region_layer_map/inception_v4.json +6 -0
- brainscore_vision/models/inception_v4/test.py +8 -0
- brainscore_vision/models/mobilenet_v2_0_5_192/__init__.py +7 -0
- brainscore_vision/models/mobilenet_v2_0_5_192/model.py +83 -0
- brainscore_vision/models/mobilenet_v2_0_5_192/region_layer_map/mobilenet_v2_0_5_192.json +6 -0
- brainscore_vision/models/mobilenet_v2_0_5_192/requirements.txt +2 -0
- brainscore_vision/models/mobilenet_v2_0_5_192/test.py +8 -0
- brainscore_vision/models/mobilenet_v2_0_5_224/__init__.py +7 -0
- brainscore_vision/models/mobilenet_v2_0_5_224/model.py +73 -0
- brainscore_vision/models/mobilenet_v2_0_5_224/region_layer_map/mobilenet_v2_0_5_224.json +6 -0
- brainscore_vision/models/mobilenet_v2_0_5_224/requirements.txt +2 -0
- brainscore_vision/models/mobilenet_v2_0_5_224/test.py +9 -0
- brainscore_vision/models/mobilenet_v2_0_75_160/__init__.py +7 -0
- brainscore_vision/models/mobilenet_v2_0_75_160/model.py +74 -0
- brainscore_vision/models/mobilenet_v2_0_75_160/region_layer_map/mobilenet_v2_0_75_160.json +6 -0
- brainscore_vision/models/mobilenet_v2_0_75_160/requirements.txt +2 -0
- brainscore_vision/models/mobilenet_v2_0_75_160/test.py +8 -0
- brainscore_vision/models/mobilenet_v2_0_75_192/__init__.py +7 -0
- brainscore_vision/models/mobilenet_v2_0_75_192/model.py +72 -0
- brainscore_vision/models/mobilenet_v2_0_75_192/region_layer_map/mobilenet_v2_0_75_192.json +6 -0
- brainscore_vision/models/mobilenet_v2_0_75_192/requirements.txt +2 -0
- brainscore_vision/models/mobilenet_v2_0_75_192/test.py +9 -0
- brainscore_vision/models/mobilenet_v2_0_75_224/__init__.py +7 -0
- brainscore_vision/models/mobilenet_v2_0_75_224/model.py +73 -0
- brainscore_vision/models/mobilenet_v2_0_75_224/region_layer_map/mobilenet_v2_0_75_224.json +6 -0
- brainscore_vision/models/mobilenet_v2_0_75_224/requirements.txt +2 -0
- brainscore_vision/models/mobilenet_v2_0_75_224/test.py +8 -0
- brainscore_vision/models/mobilenet_v2_1_0_128/__init__.py +7 -0
- brainscore_vision/models/mobilenet_v2_1_0_128/model.py +73 -0
- brainscore_vision/models/mobilenet_v2_1_0_128/region_layer_map/mobilenet_v2_1_0_128.json +6 -0
- brainscore_vision/models/mobilenet_v2_1_0_128/requirements.txt +2 -0
- brainscore_vision/models/mobilenet_v2_1_0_128/test.py +8 -0
- brainscore_vision/models/mobilenet_v2_1_0_160/__init__.py +7 -0
- brainscore_vision/models/mobilenet_v2_1_0_160/model.py +73 -0
- brainscore_vision/models/mobilenet_v2_1_0_160/region_layer_map/mobilenet_v2_1_0_160.json +6 -0
- brainscore_vision/models/mobilenet_v2_1_0_160/requirements.txt +2 -0
- brainscore_vision/models/mobilenet_v2_1_0_160/test.py +8 -0
- brainscore_vision/models/mobilenet_v2_1_0_192/__init__.py +7 -0
- brainscore_vision/models/mobilenet_v2_1_0_192/model.py +73 -0
- brainscore_vision/models/mobilenet_v2_1_0_192/region_layer_map/mobilenet_v2_1_0_192.json +6 -0
- brainscore_vision/models/mobilenet_v2_1_0_192/requirements.txt +2 -0
- brainscore_vision/models/mobilenet_v2_1_0_192/test.py +8 -0
- brainscore_vision/models/{pnasnet_large_pytorch → mobilenet_v2_1_0_224}/__init__.py +3 -3
- brainscore_vision/models/mobilenet_v2_1_0_224/model.py +60 -0
- brainscore_vision/models/mobilenet_v2_1_0_224/region_layer_map/mobilenet_v2_1_0_224.json +6 -0
- brainscore_vision/models/mobilenet_v2_1_0_224/test.py +8 -0
- brainscore_vision/models/mobilenet_v2_1_3_224/__init__.py +7 -0
- brainscore_vision/models/mobilenet_v2_1_3_224/model.py +73 -0
- brainscore_vision/models/mobilenet_v2_1_3_224/region_layer_map/mobilenet_v2_1_3_224.json +6 -0
- brainscore_vision/models/mobilenet_v2_1_3_224/requirements.txt +2 -0
- brainscore_vision/models/mobilenet_v2_1_3_224/test.py +8 -0
- brainscore_vision/models/mobilenet_v2_1_4_224/__init__.py +7 -0
- brainscore_vision/models/{mobilenet_v2_1_4_224_pytorch → mobilenet_v2_1_4_224}/model.py +3 -3
- brainscore_vision/models/mobilenet_v2_1_4_224/region_layer_map/mobilenet_v2_1_4_224.json +6 -0
- brainscore_vision/models/mobilenet_v2_1_4_224/requirements.txt +3 -0
- brainscore_vision/models/mobilenet_v2_1_4_224/test.py +8 -0
- brainscore_vision/models/nasnet_large/__init__.py +7 -0
- brainscore_vision/models/nasnet_large/model.py +60 -0
- brainscore_vision/models/nasnet_large/region_layer_map/nasnet_large.json +6 -0
- brainscore_vision/models/nasnet_large/test.py +8 -0
- brainscore_vision/models/nasnet_mobile/__init__.py +7 -0
- brainscore_vision/models/nasnet_mobile/model.py +685 -0
- brainscore_vision/models/nasnet_mobile/region_layer_map/nasnet_mobile.json +6 -0
- brainscore_vision/models/nasnet_mobile/requirements.txt +1 -0
- brainscore_vision/models/nasnet_mobile/test.py +8 -0
- brainscore_vision/models/omnivore_swinB/__init__.py +7 -0
- brainscore_vision/models/omnivore_swinB/model.py +79 -0
- brainscore_vision/models/omnivore_swinB/region_layer_map/omnivore_swinB.json +1 -0
- brainscore_vision/models/omnivore_swinB/requirements.txt +5 -0
- brainscore_vision/models/omnivore_swinB/test.py +9 -0
- brainscore_vision/models/omnivore_swinS/__init__.py +7 -0
- brainscore_vision/models/omnivore_swinS/model.py +79 -0
- brainscore_vision/models/omnivore_swinS/region_layer_map/omnivore_swinS.json +1 -0
- brainscore_vision/models/omnivore_swinS/requirements.txt +7 -0
- brainscore_vision/models/omnivore_swinS/test.py +9 -0
- brainscore_vision/models/pnasnet_large/__init__.py +7 -0
- brainscore_vision/models/{pnasnet_large_pytorch → pnasnet_large}/model.py +6 -10
- brainscore_vision/models/pnasnet_large/region_layer_map/pnasnet_large.json +6 -0
- brainscore_vision/models/pnasnet_large/requirements.txt +3 -0
- brainscore_vision/models/pnasnet_large/test.py +8 -0
- brainscore_vision/models/resnet50_SIN/__init__.py +7 -0
- brainscore_vision/models/resnet50_SIN/model.py +63 -0
- brainscore_vision/models/resnet50_SIN/region_layer_map/resnet50-SIN.json +6 -0
- brainscore_vision/models/resnet50_SIN/requirements.txt +1 -0
- brainscore_vision/models/resnet50_SIN/test.py +9 -0
- brainscore_vision/models/resnet50_SIN_IN/__init__.py +7 -0
- brainscore_vision/models/resnet50_SIN_IN/model.py +65 -0
- brainscore_vision/models/resnet50_SIN_IN/region_layer_map/resnet50-SIN_IN.json +6 -0
- brainscore_vision/models/resnet50_SIN_IN/requirements.txt +2 -0
- brainscore_vision/models/resnet50_SIN_IN/test.py +9 -0
- brainscore_vision/models/resnet50_SIN_IN_IN/__init__.py +7 -0
- brainscore_vision/models/resnet50_SIN_IN_IN/model.py +65 -0
- brainscore_vision/models/resnet50_SIN_IN_IN/region_layer_map/resnet50-SIN_IN_IN.json +6 -0
- brainscore_vision/models/resnet50_SIN_IN_IN/requirements.txt +2 -0
- brainscore_vision/models/resnet50_SIN_IN_IN/test.py +9 -0
- brainscore_vision/models/resnet50_VITO_8deg_cc/__init__.py +9 -0
- brainscore_vision/models/resnet50_VITO_8deg_cc/helpers/resnet.py +1061 -0
- brainscore_vision/models/resnet50_VITO_8deg_cc/helpers/spatialattn.py +50 -0
- brainscore_vision/models/resnet50_VITO_8deg_cc/model.py +72 -0
- brainscore_vision/models/resnet50_VITO_8deg_cc/region_layer_map/resnet50-VITO-8deg-cc.json +6 -0
- brainscore_vision/models/resnet50_VITO_8deg_cc/requirements.txt +3 -0
- brainscore_vision/models/resnet50_VITO_8deg_cc/test.py +8 -0
- brainscore_vision/models/resnet50_barlow/__init__.py +7 -0
- brainscore_vision/models/resnet50_barlow/model.py +53 -0
- brainscore_vision/models/resnet50_barlow/region_layer_map/resnet50-barlow.json +1 -0
- brainscore_vision/models/resnet50_barlow/requirements.txt +1 -0
- brainscore_vision/models/resnet50_barlow/test.py +9 -0
- brainscore_vision/models/resnet50_finetune_cutmix_AVGe2e3_robust_linf8255_e0_247x234/__init__.py +6 -0
- brainscore_vision/models/resnet50_finetune_cutmix_AVGe2e3_robust_linf8255_e0_247x234/model.py +128 -0
- brainscore_vision/models/resnet50_finetune_cutmix_AVGe2e3_robust_linf8255_e0_247x234/region_layer_map/resnet50_finetune_cutmix_AVGe2e3_robust_linf8255_e0_247x234.json +1 -0
- brainscore_vision/models/resnet50_finetune_cutmix_AVGe2e3_robust_linf8255_e0_247x234/requirements.txt +5 -0
- brainscore_vision/models/resnet50_finetune_cutmix_AVGe2e3_robust_linf8255_e0_247x234/test.py +7 -0
- brainscore_vision/models/resnet50_moclr8deg/__init__.py +11 -0
- brainscore_vision/models/resnet50_moclr8deg/helpers/helpers.py +496 -0
- brainscore_vision/models/resnet50_moclr8deg/model.py +45 -0
- brainscore_vision/models/resnet50_moclr8deg/region_layer_map/resnet50-moclr8deg.json +6 -0
- brainscore_vision/models/resnet50_moclr8deg/requirements.txt +3 -0
- brainscore_vision/models/resnet50_moclr8deg/test.py +8 -0
- brainscore_vision/models/resnet50_robust_l2_eps1/__init__.py +9 -0
- brainscore_vision/models/resnet50_robust_l2_eps1/model.py +72 -0
- brainscore_vision/models/resnet50_robust_l2_eps1/region_layer_map/resnet50_robust_l2_eps1.json +1 -0
- brainscore_vision/models/resnet50_robust_l2_eps1/requirements.txt +2 -0
- brainscore_vision/models/resnet50_robust_l2_eps1/test.py +8 -0
- brainscore_vision/models/resnet50_robust_l2_eps3/__init__.py +8 -0
- brainscore_vision/models/resnet50_robust_l2_eps3/model.py +72 -0
- brainscore_vision/models/resnet50_robust_l2_eps3/region_layer_map/resnet50_robust_l2_eps3.json +1 -0
- brainscore_vision/models/resnet50_robust_l2_eps3/requirements.txt +2 -0
- brainscore_vision/models/resnet50_robust_l2_eps3/test.py +8 -0
- brainscore_vision/models/resnet50_sup/__init__.py +5 -0
- brainscore_vision/models/resnet50_sup/model.py +55 -0
- brainscore_vision/models/resnet50_sup/region_layer_map/resnet50-sup.json +1 -0
- brainscore_vision/models/resnet50_sup/requirements.txt +1 -0
- brainscore_vision/models/resnet50_sup/test.py +8 -0
- brainscore_vision/models/resnet50_vicreg/__init__.py +7 -0
- brainscore_vision/models/resnet50_vicreg/model.py +62 -0
- brainscore_vision/models/resnet50_vicreg/region_layer_map/resnet50-vicreg.json +1 -0
- brainscore_vision/models/resnet50_vicreg/requirements.txt +1 -0
- brainscore_vision/models/resnet50_vicreg/test.py +9 -0
- brainscore_vision/models/resnet50_vicregl0p75/__init__.py +5 -0
- brainscore_vision/models/resnet50_vicregl0p75/model.py +80 -0
- brainscore_vision/models/resnet50_vicregl0p75/region_layer_map/resnet50-vicregl0p75.json +1 -0
- brainscore_vision/models/resnet50_vicregl0p75/test.py +9 -0
- brainscore_vision/models/resnet50_vicregl0p9/__init__.py +5 -0
- brainscore_vision/models/resnet50_vicregl0p9/model.py +85 -0
- brainscore_vision/models/resnet50_vicregl0p9/region_layer_map/resnet50-vicregl0p9.json +1 -0
- brainscore_vision/models/resnet50_vicregl0p9/requirements.txt +3 -0
- brainscore_vision/models/resnet50_vicregl0p9/test.py +9 -0
- brainscore_vision/models/resnet50_vitoimagevidnet8/__init__.py +11 -0
- brainscore_vision/models/resnet50_vitoimagevidnet8/helpers/helpers.py +496 -0
- brainscore_vision/models/resnet50_vitoimagevidnet8/model.py +45 -0
- brainscore_vision/models/resnet50_vitoimagevidnet8/region_layer_map/resnet50-vitoimagevidnet8.json +6 -0
- brainscore_vision/models/resnet50_vitoimagevidnet8/requirements.txt +3 -0
- brainscore_vision/models/resnet50_vitoimagevidnet8/test.py +8 -0
- brainscore_vision/models/resnet_101_v1/__init__.py +5 -0
- brainscore_vision/models/resnet_101_v1/model.py +42 -0
- brainscore_vision/models/resnet_101_v1/region_layer_map/resnet_101_v1.json +6 -0
- brainscore_vision/models/resnet_101_v1/requirements.txt +1 -0
- brainscore_vision/models/resnet_101_v1/test.py +8 -0
- brainscore_vision/models/resnet_101_v2/__init__.py +8 -0
- brainscore_vision/models/resnet_101_v2/model.py +33 -0
- brainscore_vision/models/resnet_101_v2/region_layer_map/resnet_101_v2.json +6 -0
- brainscore_vision/models/resnet_101_v2/requirements.txt +2 -0
- brainscore_vision/models/resnet_101_v2/test.py +8 -0
- brainscore_vision/models/resnet_152_v1/__init__.py +5 -0
- brainscore_vision/models/resnet_152_v1/model.py +42 -0
- brainscore_vision/models/resnet_152_v1/region_layer_map/resnet_152_v1.json +6 -0
- brainscore_vision/models/resnet_152_v1/requirements.txt +1 -0
- brainscore_vision/models/resnet_152_v1/test.py +8 -0
- brainscore_vision/models/resnet_152_v2/__init__.py +7 -0
- brainscore_vision/models/{resnet_152_v2_pytorch → resnet_152_v2}/model.py +9 -11
- brainscore_vision/models/resnet_152_v2/region_layer_map/resnet_152_v2.json +6 -0
- brainscore_vision/models/resnet_152_v2/requirements.txt +2 -0
- brainscore_vision/models/resnet_152_v2/test.py +8 -0
- brainscore_vision/models/resnet_18_test_m/__init__.py +9 -0
- brainscore_vision/models/resnet_18_test_m/helpers/resnet.py +586 -0
- brainscore_vision/models/resnet_18_test_m/model.py +80 -0
- brainscore_vision/models/resnet_18_test_m/region_layer_map/resnet-18_test_m.json +1 -0
- brainscore_vision/models/resnet_18_test_m/requirements.txt +2 -0
- brainscore_vision/models/resnet_18_test_m/test.py +8 -0
- brainscore_vision/models/resnet_50_2/__init__.py +9 -0
- brainscore_vision/models/resnet_50_2/evnet/backends.py +109 -0
- brainscore_vision/models/resnet_50_2/evnet/evnet.py +147 -0
- brainscore_vision/models/resnet_50_2/evnet/modules.py +308 -0
- brainscore_vision/models/resnet_50_2/evnet/params.py +326 -0
- brainscore_vision/models/resnet_50_2/evnet/utils.py +142 -0
- brainscore_vision/models/resnet_50_2/model.py +46 -0
- brainscore_vision/models/resnet_50_2/region_layer_map/resnet_50_2.json +6 -0
- brainscore_vision/models/resnet_50_2/requirements.txt +4 -0
- brainscore_vision/models/resnet_50_2/test.py +8 -0
- brainscore_vision/models/resnet_50_robust/model.py +2 -2
- brainscore_vision/models/resnet_50_robust/region_layer_map/resnet-50-robust.json +1 -0
- brainscore_vision/models/resnet_50_v1/__init__.py +5 -0
- brainscore_vision/models/resnet_50_v1/model.py +42 -0
- brainscore_vision/models/resnet_50_v1/region_layer_map/resnet_50_v1.json +6 -0
- brainscore_vision/models/resnet_50_v1/requirements.txt +1 -0
- brainscore_vision/models/resnet_50_v1/test.py +8 -0
- brainscore_vision/models/resnet_50_v2/__init__.py +8 -0
- brainscore_vision/models/resnet_50_v2/model.py +33 -0
- brainscore_vision/models/resnet_50_v2/region_layer_map/resnet_50_v2.json +6 -0
- brainscore_vision/models/resnet_50_v2/requirements.txt +2 -0
- brainscore_vision/models/resnet_50_v2/test.py +8 -0
- brainscore_vision/models/resnet_SIN_IN_FT_IN/__init__.py +5 -0
- brainscore_vision/models/resnet_SIN_IN_FT_IN/model.py +79 -0
- brainscore_vision/models/resnet_SIN_IN_FT_IN/region_layer_map/resnet_SIN_IN_FT_IN.json +1 -0
- brainscore_vision/models/resnet_SIN_IN_FT_IN/requirements.txt +2 -0
- brainscore_vision/models/resnet_SIN_IN_FT_IN/test.py +8 -0
- brainscore_vision/models/sBarlow_lmda_0/__init__.py +9 -0
- brainscore_vision/models/sBarlow_lmda_0/model.py +64 -0
- brainscore_vision/models/sBarlow_lmda_0/region_layer_map/sBarlow_lmda_0.json +6 -0
- brainscore_vision/models/sBarlow_lmda_0/setup.py +25 -0
- brainscore_vision/models/sBarlow_lmda_0/test.py +1 -0
- brainscore_vision/models/sBarlow_lmda_01/__init__.py +9 -0
- brainscore_vision/models/sBarlow_lmda_01/model.py +64 -0
- brainscore_vision/models/sBarlow_lmda_01/region_layer_map/sBarlow_lmda_01.json +6 -0
- brainscore_vision/models/sBarlow_lmda_01/setup.py +25 -0
- brainscore_vision/models/sBarlow_lmda_01/test.py +1 -0
- brainscore_vision/models/sBarlow_lmda_1/__init__.py +9 -0
- brainscore_vision/models/sBarlow_lmda_1/model.py +64 -0
- brainscore_vision/models/sBarlow_lmda_1/region_layer_map/sBarlow_lmda_1.json +6 -0
- brainscore_vision/models/sBarlow_lmda_1/setup.py +25 -0
- brainscore_vision/models/sBarlow_lmda_1/test.py +1 -0
- brainscore_vision/models/sBarlow_lmda_2/__init__.py +9 -0
- brainscore_vision/models/sBarlow_lmda_2/model.py +64 -0
- brainscore_vision/models/sBarlow_lmda_2/region_layer_map/sBarlow_lmda_2.json +6 -0
- brainscore_vision/models/sBarlow_lmda_2/setup.py +25 -0
- brainscore_vision/models/sBarlow_lmda_2/test.py +1 -0
- brainscore_vision/models/sBarlow_lmda_8/__init__.py +9 -0
- brainscore_vision/models/sBarlow_lmda_8/model.py +64 -0
- brainscore_vision/models/sBarlow_lmda_8/region_layer_map/sBarlow_lmda_8.json +6 -0
- brainscore_vision/models/sBarlow_lmda_8/setup.py +25 -0
- brainscore_vision/models/sBarlow_lmda_8/test.py +1 -0
- brainscore_vision/models/scsBarlow_lmda_1/__init__.py +9 -0
- brainscore_vision/models/scsBarlow_lmda_1/model.py +64 -0
- brainscore_vision/models/scsBarlow_lmda_1/region_layer_map/scsBarlow_lmda_1.json +6 -0
- brainscore_vision/models/scsBarlow_lmda_1/setup.py +25 -0
- brainscore_vision/models/scsBarlow_lmda_1/test.py +1 -0
- brainscore_vision/models/scsBarlow_lmda_2/__init__.py +9 -0
- brainscore_vision/models/scsBarlow_lmda_2/model.py +64 -0
- brainscore_vision/models/scsBarlow_lmda_2/region_layer_map/scsBarlow_lmda_2.json +6 -0
- brainscore_vision/models/scsBarlow_lmda_2/setup.py +25 -0
- brainscore_vision/models/scsBarlow_lmda_2/test.py +1 -0
- brainscore_vision/models/scsBarlow_lmda_4/__init__.py +9 -0
- brainscore_vision/models/scsBarlow_lmda_4/model.py +64 -0
- brainscore_vision/models/scsBarlow_lmda_4/region_layer_map/scsBarlow_lmda_4.json +6 -0
- brainscore_vision/models/scsBarlow_lmda_4/setup.py +25 -0
- brainscore_vision/models/scsBarlow_lmda_4/test.py +1 -0
- brainscore_vision/models/shufflenet_v2_x1_0/__init__.py +7 -0
- brainscore_vision/models/shufflenet_v2_x1_0/model.py +52 -0
- brainscore_vision/models/shufflenet_v2_x1_0/region_layer_map/shufflenet_v2_x1_0.json +1 -0
- brainscore_vision/models/shufflenet_v2_x1_0/requirements.txt +2 -0
- brainscore_vision/models/shufflenet_v2_x1_0/test.py +9 -0
- brainscore_vision/models/timm_models/__init__.py +193 -0
- brainscore_vision/models/timm_models/model.py +90 -0
- brainscore_vision/models/timm_models/model_configs.json +464 -0
- brainscore_vision/models/timm_models/requirements.txt +3 -0
- brainscore_vision/models/timm_models/test.py +0 -0
- brainscore_vision/models/vgg_16/__init__.py +7 -0
- brainscore_vision/models/vgg_16/model.py +52 -0
- brainscore_vision/models/vgg_16/region_layer_map/vgg_16.json +6 -0
- brainscore_vision/models/vgg_16/requirements.txt +1 -0
- brainscore_vision/models/vgg_16/test.py +8 -0
- brainscore_vision/models/vgg_19/__init__.py +7 -0
- brainscore_vision/models/vgg_19/model.py +52 -0
- brainscore_vision/models/vgg_19/region_layer_map/vgg_19.json +1 -0
- brainscore_vision/models/vgg_19/requirements.txt +1 -0
- brainscore_vision/models/vgg_19/test.py +8 -0
- brainscore_vision/models/vonegrcnn_47e/__init__.py +5 -0
- brainscore_vision/models/vonegrcnn_47e/model.py +622 -0
- brainscore_vision/models/vonegrcnn_47e/region_layer_map/vonegrcnn_47e.json +6 -0
- brainscore_vision/models/vonegrcnn_47e/requirements.txt +0 -0
- brainscore_vision/models/vonegrcnn_47e/test.py +8 -0
- brainscore_vision/models/vonegrcnn_52e_full/__init__.py +5 -0
- brainscore_vision/models/vonegrcnn_52e_full/model.py +623 -0
- brainscore_vision/models/vonegrcnn_52e_full/region_layer_map/vonegrcnn_52e_full.json +6 -0
- brainscore_vision/models/vonegrcnn_52e_full/requirements.txt +4 -0
- brainscore_vision/models/vonegrcnn_52e_full/test.py +8 -0
- brainscore_vision/models/vonegrcnn_62e_nobn/__init__.py +7 -0
- brainscore_vision/models/vonegrcnn_62e_nobn/helpers/vongrcnn_helpers.py +544 -0
- brainscore_vision/models/vonegrcnn_62e_nobn/model.py +122 -0
- brainscore_vision/models/vonegrcnn_62e_nobn/region_layer_map/vonegrcnn_62e_nobn.json +6 -0
- brainscore_vision/models/vonegrcnn_62e_nobn/requirements.txt +3 -0
- brainscore_vision/models/vonegrcnn_62e_nobn/test.py +8 -0
- brainscore_vision/models/voneresnet_50/__init__.py +7 -0
- brainscore_vision/models/voneresnet_50/model.py +37 -0
- brainscore_vision/models/voneresnet_50/region_layer_map/voneresnet-50.json +6 -0
- brainscore_vision/models/voneresnet_50/requirements.txt +1 -0
- brainscore_vision/models/voneresnet_50/test.py +8 -0
- brainscore_vision/models/voneresnet_50_1/__init__.py +11 -0
- brainscore_vision/models/voneresnet_50_1/evnet/backends.py +109 -0
- brainscore_vision/models/voneresnet_50_1/evnet/evnet.py +147 -0
- brainscore_vision/models/voneresnet_50_1/evnet/modules.py +308 -0
- brainscore_vision/models/voneresnet_50_1/evnet/params.py +326 -0
- brainscore_vision/models/voneresnet_50_1/evnet/utils.py +142 -0
- brainscore_vision/models/voneresnet_50_1/model.py +68 -0
- brainscore_vision/models/voneresnet_50_1/requirements.txt +5 -0
- brainscore_vision/models/voneresnet_50_1/test.py +7 -0
- brainscore_vision/models/voneresnet_50_3/__init__.py +11 -0
- brainscore_vision/models/voneresnet_50_3/evnet/backends.py +109 -0
- brainscore_vision/models/voneresnet_50_3/evnet/evnet.py +147 -0
- brainscore_vision/models/voneresnet_50_3/evnet/modules.py +308 -0
- brainscore_vision/models/voneresnet_50_3/evnet/params.py +326 -0
- brainscore_vision/models/voneresnet_50_3/evnet/utils.py +142 -0
- brainscore_vision/models/voneresnet_50_3/model.py +66 -0
- brainscore_vision/models/voneresnet_50_3/requirements.txt +4 -0
- brainscore_vision/models/voneresnet_50_3/test.py +7 -0
- brainscore_vision/models/voneresnet_50_no_weight/__init__.py +11 -0
- brainscore_vision/models/voneresnet_50_no_weight/evnet/backends.py +109 -0
- brainscore_vision/models/voneresnet_50_no_weight/evnet/evnet.py +147 -0
- brainscore_vision/models/voneresnet_50_no_weight/evnet/modules.py +308 -0
- brainscore_vision/models/voneresnet_50_no_weight/evnet/params.py +326 -0
- brainscore_vision/models/voneresnet_50_no_weight/evnet/utils.py +142 -0
- brainscore_vision/models/voneresnet_50_no_weight/model.py +56 -0
- brainscore_vision/models/voneresnet_50_no_weight/requirements.txt +4 -0
- brainscore_vision/models/voneresnet_50_no_weight/test.py +7 -0
- brainscore_vision/models/voneresnet_50_non_stochastic/model.py +2 -2
- brainscore_vision/models/voneresnet_50_robust/__init__.py +7 -0
- brainscore_vision/models/voneresnet_50_robust/model.py +50 -0
- brainscore_vision/models/voneresnet_50_robust/region_layer_map/voneresnet-50-robust.json +6 -0
- brainscore_vision/models/voneresnet_50_robust/requirements.txt +1 -0
- brainscore_vision/models/voneresnet_50_robust/test.py +8 -0
- brainscore_vision/models/xception/__init__.py +7 -0
- brainscore_vision/models/xception/model.py +64 -0
- brainscore_vision/models/xception/region_layer_map/xception.json +6 -0
- brainscore_vision/models/xception/requirements.txt +2 -0
- brainscore_vision/models/xception/test.py +8 -0
- brainscore_vision/models/yudixie_resnet50_250117_0/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_0/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_0/region_layer_map/yudixie_resnet50_distance_reg_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_0/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_0/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_1/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_1/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_1/region_layer_map/yudixie_resnet50_translation_reg_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_1/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_1/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_10/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_10/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_10/region_layer_map/yudixie_resnet50_imagenet1kpret_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_10/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_10/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_11/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_11/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_11/region_layer_map/yudixie_resnet50_random_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_11/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_11/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_2/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_2/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_2/region_layer_map/yudixie_resnet50_rotation_reg_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_2/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_2/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_3/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_3/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_3/region_layer_map/yudixie_resnet50_distance_translation_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_3/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_3/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_4/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_4/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_4/region_layer_map/yudixie_resnet50_distance_rotation_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_4/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_4/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_5/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_5/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_5/region_layer_map/yudixie_resnet50_translation_rotation_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_5/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_5/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_6/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_6/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_6/region_layer_map/yudixie_resnet50_distance_translation_rotation_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_6/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_6/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_7/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_7/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_7/region_layer_map/yudixie_resnet50_category_class_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_7/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_7/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_8/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_8/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_8/region_layer_map/yudixie_resnet50_object_class_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_8/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_8/test.py +1 -0
- brainscore_vision/models/yudixie_resnet50_250117_9/__init__.py +11 -0
- brainscore_vision/models/yudixie_resnet50_250117_9/model.py +60 -0
- brainscore_vision/models/yudixie_resnet50_250117_9/region_layer_map/yudixie_resnet50_cat_obj_class_all_latents_0_240908.json +6 -0
- brainscore_vision/models/yudixie_resnet50_250117_9/setup.py +25 -0
- brainscore_vision/models/yudixie_resnet50_250117_9/test.py +1 -0
- brainscore_vision/submission/actions_helpers.py +2 -3
- {brainscore_vision-2.2.3.dist-info → brainscore_vision-2.2.5.dist-info}/METADATA +6 -6
- {brainscore_vision-2.2.3.dist-info → brainscore_vision-2.2.5.dist-info}/RECORD +714 -130
- {brainscore_vision-2.2.3.dist-info → brainscore_vision-2.2.5.dist-info}/WHEEL +1 -1
- docs/source/index.rst +1 -0
- docs/source/modules/submission.rst +1 -1
- docs/source/modules/version_bumping.rst +43 -0
- tests/test_submission/test_actions_helpers.py +2 -6
- brainscore_vision/models/densenet_201_pytorch/test.py +0 -8
- brainscore_vision/models/inception_v3_pytorch/test.py +0 -8
- brainscore_vision/models/inception_v4_pytorch/test.py +0 -8
- brainscore_vision/models/mobilenet_v2_1_4_224_pytorch/__init__.py +0 -7
- brainscore_vision/models/mobilenet_v2_1_4_224_pytorch/test.py +0 -8
- brainscore_vision/models/pnasnet_large_pytorch/test.py +0 -8
- brainscore_vision/models/resnet_152_v2_pytorch/__init__.py +0 -7
- brainscore_vision/models/resnet_152_v2_pytorch/test.py +0 -8
- /brainscore_vision/models/{densenet_201_pytorch → densenet_201}/requirements.txt +0 -0
- /brainscore_vision/models/{inception_v3_pytorch → inception_v3}/requirements.txt +0 -0
- /brainscore_vision/models/{inception_v4_pytorch → inception_v4}/requirements.txt +0 -0
- /brainscore_vision/models/{mobilenet_v2_1_4_224_pytorch → mobilenet_v2_1_0_224}/requirements.txt +0 -0
- /brainscore_vision/models/{pnasnet_large_pytorch → nasnet_large}/requirements.txt +0 -0
- /brainscore_vision/models/{resnet_152_v2_pytorch → resnet50_vicregl0p75}/requirements.txt +0 -0
- {brainscore_vision-2.2.3.dist-info → brainscore_vision-2.2.5.dist-info}/LICENSE +0 -0
- {brainscore_vision-2.2.3.dist-info → brainscore_vision-2.2.5.dist-info}/top_level.txt +0 -0
@@ -6,17 +6,17 @@ from brainscore_vision.data_helpers.s3 import load_assembly_from_s3, load_stimul
|
|
6
6
|
# normal distortion:
|
7
7
|
stimulus_set_registry['Baker2022_normal_distortion'] = lambda: load_stimulus_set_from_s3(
|
8
8
|
identifier='Baker2022_normal_distortion',
|
9
|
-
bucket="brainio-brainscore",
|
9
|
+
bucket="brainscore-storage/brainio-brainscore",
|
10
10
|
csv_sha1="17d4db7458a29a787d12bb29c34e91daef1872bf",
|
11
11
|
zip_sha1="2c726abaf081c8a9828269a559222f8c6eea0e4f",
|
12
|
-
csv_version_id="
|
13
|
-
zip_version_id="
|
12
|
+
csv_version_id="null",
|
13
|
+
zip_version_id="null")
|
14
14
|
|
15
15
|
data_registry['Baker2022_normal_distortion'] = lambda: load_assembly_from_s3(
|
16
16
|
identifier='Baker2022_normal_distortion',
|
17
|
-
version_id="
|
17
|
+
version_id="null",
|
18
18
|
sha1="46c79a48bf2dbd995a9305d8fbc03a134a852e17",
|
19
|
-
bucket="brainio-brainscore",
|
19
|
+
bucket="brainscore-storage/brainio-brainscore",
|
20
20
|
cls=BehavioralAssembly,
|
21
21
|
stimulus_set_loader=lambda: load_stimulus_set('Baker2022_normal_distortion'),
|
22
22
|
)
|
@@ -24,17 +24,17 @@ data_registry['Baker2022_normal_distortion'] = lambda: load_assembly_from_s3(
|
|
24
24
|
# inverted distortion:
|
25
25
|
stimulus_set_registry['Baker2022_inverted_distortion'] = lambda: load_stimulus_set_from_s3(
|
26
26
|
identifier='Baker2022_inverted_distortion',
|
27
|
-
bucket="brainio-brainscore",
|
27
|
+
bucket="brainscore-storage/brainio-brainscore",
|
28
28
|
csv_sha1="91e452e4651024c2b382694edfcbc7bdc6c3189b",
|
29
29
|
zip_sha1="4740a096af994c2232350469c664e53796f17a05",
|
30
|
-
csv_version_id="
|
31
|
-
zip_version_id="
|
30
|
+
csv_version_id="null",
|
31
|
+
zip_version_id="null")
|
32
32
|
|
33
33
|
data_registry['Baker2022_inverted_distortion'] = lambda: load_assembly_from_s3(
|
34
34
|
identifier='Baker2022_inverted_distortion',
|
35
|
-
version_id="
|
35
|
+
version_id="null",
|
36
36
|
sha1="b76fb57b25a58ca68db78d188fd0a783e1dcaf73",
|
37
|
-
bucket="brainio-brainscore",
|
37
|
+
bucket="brainscore-storage/brainio-brainscore",
|
38
38
|
cls=BehavioralAssembly,
|
39
39
|
stimulus_set_loader=lambda: load_stimulus_set('Baker2022_inverted_distortion'),
|
40
40
|
)
|
@@ -38,6 +38,6 @@ assembly = BehavioralAssembly(all_subjects['RSP'],
|
|
38
38
|
assembly.name = 'Baker2022_inverted_distortion'
|
39
39
|
|
40
40
|
# upload to S3
|
41
|
-
package_data_assembly('brainio_brainscore', assembly, assembly_identifier=assembly.name,
|
41
|
+
package_data_assembly('brainscore-storage/brainio_brainscore', assembly, assembly_identifier=assembly.name,
|
42
42
|
stimulus_set_identifier='Baker2022_inverted_distortion',
|
43
|
-
assembly_class_name="BehavioralAssembly", bucket_name="
|
43
|
+
assembly_class_name="BehavioralAssembly", bucket_name="brainscore-storage/brainio_brainscore")
|
@@ -77,5 +77,5 @@ stimuli.stimulus_paths = image_paths
|
|
77
77
|
stimuli.name = 'Baker2022_inverted_distortion'
|
78
78
|
|
79
79
|
# upload to S3
|
80
|
-
package_stimulus_set("brainio_brainscore", stimuli, stimulus_set_identifier=stimuli.name,
|
81
|
-
bucket_name="
|
80
|
+
package_stimulus_set("brainscore-storage/brainio_brainscore", stimuli, stimulus_set_identifier=stimuli.name,
|
81
|
+
bucket_name="brainscore-storage/brainio_brainscore")
|
@@ -41,6 +41,6 @@ assembly.name = 'Baker2022_normal_distortion'
|
|
41
41
|
|
42
42
|
|
43
43
|
# upload to S3
|
44
|
-
package_data_assembly('brainio_brainscore', assembly, assembly_identifier=assembly.name,
|
44
|
+
package_data_assembly('brainscore-storage/brainio_brainscore', assembly, assembly_identifier=assembly.name,
|
45
45
|
stimulus_set_identifier='Baker2022_normal_distortion',
|
46
|
-
assembly_class_name="BehavioralAssembly", bucket_name="
|
46
|
+
assembly_class_name="BehavioralAssembly", bucket_name="brainscore-storage/brainio_brainscore")
|
@@ -90,5 +90,5 @@ stimuli = stimuli[stimuli["orientation"] == "normal"]
|
|
90
90
|
stimuli.name = 'Baker2022_normal_distortion' # give the StimulusSet an identifier name
|
91
91
|
|
92
92
|
# upload to S3
|
93
|
-
package_stimulus_set("brainio_brainscore", stimuli, stimulus_set_identifier=stimuli.name,
|
94
|
-
bucket_name="
|
93
|
+
package_stimulus_set("brainscore-storage/brainio_brainscore", stimuli, stimulus_set_identifier=stimuli.name,
|
94
|
+
bucket_name="brainscore-storage/brainio_brainscore")
|
@@ -16,8 +16,8 @@ BIBTEX = """@inproceedings{NEURIPS2019_97af07a1,
|
|
16
16
|
# stimulus set
|
17
17
|
stimulus_set_registry['BarbuMayo2019'] = lambda: load_stimulus_set_from_s3(
|
18
18
|
identifier="BarbuMayo2019",
|
19
|
-
bucket="brainio-brainscore",
|
19
|
+
bucket="brainscore-storage/brainio-brainscore",
|
20
20
|
csv_sha1="e4d8888ccb6beca28636e6698e7beb130e278e12",
|
21
21
|
zip_sha1="1365eb2a7231516806127a7d2a908343a7ac9464",
|
22
|
-
csv_version_id="
|
23
|
-
zip_version_id="
|
22
|
+
csv_version_id="null",
|
23
|
+
zip_version_id="null")
|
@@ -16,18 +16,18 @@ BIBTEX = """@article{bashivan2019neural,
|
|
16
16
|
# assemblies
|
17
17
|
data_registry['BashivanKar2019.naturalistic'] = lambda: load_assembly_from_s3(
|
18
18
|
identifier="dicarlo.BashivanKar2019.naturalistic",
|
19
|
-
version_id="
|
19
|
+
version_id="null",
|
20
20
|
sha1="1ec2f32ef800f0c6e15879d883be1d55b51b8b67",
|
21
|
-
bucket="brainio-brainscore",
|
21
|
+
bucket="brainscore-storage/brainio-brainscore",
|
22
22
|
cls=NeuronRecordingAssembly,
|
23
23
|
stimulus_set_loader=lambda: load_stimulus_set('BashivanKar2019.naturalistic'),
|
24
24
|
)
|
25
25
|
|
26
26
|
data_registry['BashivanKar2019.synthetic'] = lambda: load_assembly_from_s3(
|
27
27
|
identifier="dicarlo.BashivanKar2019.synthetic",
|
28
|
-
version_id="
|
28
|
+
version_id="null",
|
29
29
|
sha1="f687c8d26f8943dc379dbcbe94d3feb148400c6b",
|
30
|
-
bucket="brainio-brainscore",
|
30
|
+
bucket="brainscore-storage/brainio-brainscore",
|
31
31
|
cls=NeuronRecordingAssembly,
|
32
32
|
stimulus_set_loader=lambda: load_stimulus_set('BashivanKar2019.synthetic'),
|
33
33
|
)
|
@@ -36,17 +36,17 @@ data_registry['BashivanKar2019.synthetic'] = lambda: load_assembly_from_s3(
|
|
36
36
|
# naturalistic
|
37
37
|
stimulus_set_registry['BashivanKar2019.naturalistic'] = lambda: load_stimulus_set_from_s3(
|
38
38
|
identifier="BashivanKar2019.naturalistic",
|
39
|
-
bucket="brainio-brainscore",
|
39
|
+
bucket="brainscore-storage/brainio-brainscore",
|
40
40
|
csv_sha1="48ef84282552b8796142ffe7d0d2c632f8ef061a",
|
41
41
|
zip_sha1="d7b71b431cf23d435395205f1e38036a9e10acca",
|
42
|
-
csv_version_id="
|
43
|
-
zip_version_id="
|
42
|
+
csv_version_id="null",
|
43
|
+
zip_version_id="null")
|
44
44
|
|
45
45
|
# synthetic
|
46
46
|
stimulus_set_registry['BashivanKar2019.synthetic'] = lambda: load_stimulus_set_from_s3(
|
47
47
|
identifier="BashivanKar2019.synthetic",
|
48
|
-
bucket="brainio-brainscore",
|
48
|
+
bucket="brainscore-storage/brainio-brainscore",
|
49
49
|
csv_sha1="81da195e9b2a128b228fc4867e23ae6b21bd7abd",
|
50
50
|
zip_sha1="e2de33f25c5c19bcfb400055c1db399d553487e5",
|
51
|
-
csv_version_id="
|
52
|
-
zip_version_id="
|
51
|
+
csv_version_id="null",
|
52
|
+
zip_version_id="null")
|
@@ -152,9 +152,9 @@ def main():
|
|
152
152
|
assert responses_synth_concat.shape ==(21360, 233, 1)
|
153
153
|
package_data_assembly(
|
154
154
|
responses_synth_concat,
|
155
|
-
assembly_identifier=
|
155
|
+
assembly_identifier=None,
|
156
156
|
stimulus_set_identifier=stimuli_synth.identifier,
|
157
|
-
bucket_name='
|
157
|
+
bucket_name='brainscore-storage/'
|
158
158
|
)
|
159
159
|
|
160
160
|
|
@@ -5,65 +5,65 @@ from brainscore_vision.data_helpers.s3 import load_assembly_from_s3, load_stimul
|
|
5
5
|
|
6
6
|
data_registry['BMD2024.texture_1'] = lambda: load_assembly_from_s3(
|
7
7
|
identifier='BMD_2024_texture_1',
|
8
|
-
version_id='
|
8
|
+
version_id='null',
|
9
9
|
sha1='050cef2bd38fe0e0c6d55c9a4ba0b1c57550a072',
|
10
|
-
bucket="brainio-brainscore",
|
10
|
+
bucket="brainscore-storage/brainio-brainscore",
|
11
11
|
cls=BehavioralAssembly,
|
12
12
|
stimulus_set_loader=lambda: load_stimulus_set('BMD2024.texture_1'))
|
13
13
|
|
14
14
|
data_registry['BMD2024.texture_2'] = lambda: load_assembly_from_s3(
|
15
15
|
identifier='BMD_2024_texture_2',
|
16
|
-
version_id='
|
16
|
+
version_id='null',
|
17
17
|
sha1='1f9f4ee938df509c0cbeaec7fdfe0f40997da331',
|
18
|
-
bucket="brainio-brainscore",
|
18
|
+
bucket="brainscore-storage/brainio-brainscore",
|
19
19
|
cls=BehavioralAssembly,
|
20
20
|
stimulus_set_loader=lambda: load_stimulus_set('BMD2024.texture_2'))
|
21
21
|
|
22
22
|
data_registry['BMD2024.dotted_1'] = lambda: load_assembly_from_s3(
|
23
23
|
identifier='BMD_2024_dotted_1',
|
24
|
-
version_id='
|
24
|
+
version_id='null',
|
25
25
|
sha1='eb16feffe392087b4c40ef249850825f702e7911',
|
26
|
-
bucket="brainio-brainscore",
|
26
|
+
bucket="brainscore-storage/brainio-brainscore",
|
27
27
|
cls=BehavioralAssembly,
|
28
28
|
stimulus_set_loader=lambda: load_stimulus_set('BMD2024.dotted_1'))
|
29
29
|
|
30
30
|
data_registry['BMD2024.dotted_2'] = lambda: load_assembly_from_s3(
|
31
31
|
identifier='BMD_2024_dotted_2',
|
32
|
-
version_id='
|
32
|
+
version_id='null',
|
33
33
|
sha1='297833a094513b99ae434e581df09ac64cd6582f',
|
34
|
-
bucket="brainio-brainscore",
|
34
|
+
bucket="brainscore-storage/brainio-brainscore",
|
35
35
|
cls=BehavioralAssembly,
|
36
36
|
stimulus_set_loader=lambda: load_stimulus_set('BMD2024.dotted_2'))
|
37
37
|
|
38
38
|
|
39
39
|
stimulus_set_registry['BMD2024.texture_1'] = lambda: load_stimulus_set_from_s3(
|
40
40
|
identifier='BMD_2024_texture_1',
|
41
|
-
bucket="brainio-brainscore",
|
41
|
+
bucket="brainscore-storage/brainio-brainscore",
|
42
42
|
csv_sha1='395911b2933d675b98dda7bae422f11648d8e86d',
|
43
43
|
zip_sha1='cfde36c93dc9070ef5dfaa0a992c9d2420af3460',
|
44
|
-
csv_version_id='
|
45
|
-
zip_version_id='
|
44
|
+
csv_version_id='null',
|
45
|
+
zip_version_id='null')
|
46
46
|
|
47
47
|
stimulus_set_registry['BMD2024.texture_2'] = lambda: load_stimulus_set_from_s3(
|
48
48
|
identifier='BMD_2024_texture_2',
|
49
|
-
bucket="brainio-brainscore",
|
49
|
+
bucket="brainscore-storage/brainio-brainscore",
|
50
50
|
csv_sha1='98ff8e3a1ca6f632ebc2daa909804314bc1b7e31',
|
51
51
|
zip_sha1='31c9d8449b25da8ad3cb034eee04db9193027fcb',
|
52
|
-
csv_version_id='
|
53
|
-
zip_version_id='
|
52
|
+
csv_version_id='null',
|
53
|
+
zip_version_id='null')
|
54
54
|
|
55
55
|
stimulus_set_registry['BMD2024.dotted_1'] = lambda: load_stimulus_set_from_s3(
|
56
56
|
identifier='BMD_2024_dotted_1',
|
57
|
-
bucket="brainio-brainscore",
|
57
|
+
bucket="brainscore-storage/brainio-brainscore",
|
58
58
|
csv_sha1='de4214666237a0be39810ec4fefd6ec8d2a2e881',
|
59
59
|
zip_sha1='b4ab1355665b5bf3bf81b7aa6eccfd396c96bda2',
|
60
|
-
csv_version_id='
|
61
|
-
zip_version_id='
|
60
|
+
csv_version_id='null',
|
61
|
+
zip_version_id='null')
|
62
62
|
|
63
63
|
stimulus_set_registry['BMD2024.dotted_2'] = lambda: load_stimulus_set_from_s3(
|
64
64
|
identifier='BMD_2024_dotted_2',
|
65
|
-
bucket="brainio-brainscore",
|
65
|
+
bucket="brainscore-storage/brainio-brainscore",
|
66
66
|
csv_sha1='4555daa5257dee10c6c6a5625d3bb2d94452e294',
|
67
67
|
zip_sha1='20337c1fac66ed0eec16410c6801cca830e6c20c',
|
68
|
-
csv_version_id='
|
69
|
-
zip_version_id='.
|
68
|
+
csv_version_id='null',
|
69
|
+
zip_version_id='dzELAKHsBx1DKkrWAR9uteJ7K1.FtlAm')
|
@@ -86,6 +86,7 @@ if __name__ == '__main__':
|
|
86
86
|
assembly_identifier=assembly.name,
|
87
87
|
stimulus_set_identifier=assembly.name,
|
88
88
|
assembly_class_name="BehavioralAssembly",
|
89
|
-
bucket_name="brainio-brainscore")
|
89
|
+
bucket_name="brainscore-storage/brainio-brainscore")
|
90
|
+
|
90
91
|
|
91
92
|
print(prints)
|
@@ -42,7 +42,8 @@ if __name__ == '__main__':
|
|
42
42
|
prints = package_stimulus_set(catalog_name=None,
|
43
43
|
proto_stimulus_set=assembly,
|
44
44
|
stimulus_set_identifier=assembly.name,
|
45
|
-
bucket_name="brainio-brainscore")
|
45
|
+
bucket_name="brainscore-storage/brainio-brainscore")
|
46
|
+
|
46
47
|
print(prints)
|
47
48
|
|
48
49
|
|
@@ -19,18 +19,18 @@ BIBTEX = """"@article{bracci2019ventral,
|
|
19
19
|
# Human Stimulus Set
|
20
20
|
stimulus_set_registry["Bracci2019"] = lambda: load_stimulus_set_from_s3(
|
21
21
|
identifier="Bracci2019",
|
22
|
-
bucket="brainio-brainscore",
|
22
|
+
bucket="brainscore-storage/brainio-brainscore",
|
23
23
|
csv_sha1="05b1af9b8e6ed478ea49339e11b0024c2da8c35f",
|
24
24
|
zip_sha1="a79b249e758421f46ec781301cd4b498f64853ce",
|
25
|
-
csv_version_id="
|
26
|
-
zip_version_id="
|
25
|
+
csv_version_id="null",
|
26
|
+
zip_version_id="null",
|
27
27
|
)
|
28
28
|
|
29
29
|
# Human Data Assembly (brain)
|
30
30
|
data_registry["Bracci2019"] = lambda: load_assembly_from_s3(
|
31
31
|
identifier="Bracci2019",
|
32
|
-
version_id="
|
32
|
+
version_id="null",
|
33
33
|
sha1="cbec165bb20f09d0527fddba7cfbf115a396a2f3",
|
34
|
-
bucket="brainio-brainscore",
|
34
|
+
bucket="brainscore-storage/brainio-brainscore",
|
35
35
|
cls=NeuroidAssembly,
|
36
36
|
)
|
@@ -23,9 +23,9 @@ BIBTEX = """@article {Cadena201764,
|
|
23
23
|
# assembly
|
24
24
|
data_registry['Cadena2017'] = lambda: reindex(load_assembly_from_s3(
|
25
25
|
identifier="tolias.Cadena2017",
|
26
|
-
version_id="
|
26
|
+
version_id="null",
|
27
27
|
sha1="69bcaaa9370dceb0027beaa06235ef418c3d7063",
|
28
|
-
bucket="brainio-brainscore",
|
28
|
+
bucket="brainscore-storage/brainio-brainscore",
|
29
29
|
cls=NeuronRecordingAssembly,
|
30
30
|
stimulus_set_loader=lambda: load_stimulus_set('Cadena2017'),
|
31
31
|
))
|
@@ -45,8 +45,8 @@ def reindex(assembly: DataAssembly) -> DataAssembly: # make sure we have the ex
|
|
45
45
|
# stimulus set
|
46
46
|
stimulus_set_registry['Cadena2017'] = lambda: load_stimulus_set_from_s3(
|
47
47
|
identifier="Cadena2017",
|
48
|
-
bucket="brainio-brainscore",
|
48
|
+
bucket="brainscore-storage/brainio-brainscore",
|
49
49
|
csv_sha1="f55b174cc4540e5612cfba5e695324328064b051",
|
50
50
|
zip_sha1="88cc2ce3ef5e197ffd1477144a2e6a68d424ef6c",
|
51
|
-
csv_version_id="
|
52
|
-
zip_version_id="
|
51
|
+
csv_version_id="null",
|
52
|
+
zip_version_id="null")
|
@@ -20,9 +20,9 @@ abstract = {The degree to which we perceive real-world objects as similar or dis
|
|
20
20
|
# assembly
|
21
21
|
data_registry['Cichy2019'] = lambda: load_assembly_from_s3(
|
22
22
|
identifier="aru.Cichy2019",
|
23
|
-
version_id="
|
23
|
+
version_id="null",
|
24
24
|
sha1="701e63be62b642082d476244d0d91d510b3ff05d",
|
25
|
-
bucket="brainio-brainscore",
|
25
|
+
bucket="brainscore-storage/brainio-brainscore",
|
26
26
|
cls=DataAssembly,
|
27
27
|
stimulus_set_loader=lambda: load_stimulus_set('Cichy2019'),
|
28
28
|
)
|
@@ -31,8 +31,8 @@ data_registry['Cichy2019'] = lambda: load_assembly_from_s3(
|
|
31
31
|
# stimulus set
|
32
32
|
stimulus_set_registry['Cichy2019'] = lambda: load_stimulus_set_from_s3(
|
33
33
|
identifier="Cichy2019",
|
34
|
-
bucket="brainio-brainscore",
|
34
|
+
bucket="brainscore-storage/brainio-brainscore",
|
35
35
|
csv_sha1="281c4d9d0dd91a2916674638098fe94afb87d29a",
|
36
36
|
zip_sha1="d2166dd9c2720cb24bc520f5041e6830779c0240",
|
37
|
-
csv_version_id="
|
38
|
-
zip_version_id="
|
37
|
+
csv_version_id="null",
|
38
|
+
zip_version_id="null")
|
@@ -9,28 +9,28 @@ from brainscore_vision.data_helpers.s3 import (
|
|
9
9
|
# stimulus set
|
10
10
|
stimulus_set_registry['Coggan2024_behavior'] = lambda: load_stimulus_set_from_s3(
|
11
11
|
identifier="tong.Coggan2024_behavior",
|
12
|
-
bucket="brainio-brainscore",
|
12
|
+
bucket="brainscore-storage/brainio-brainscore",
|
13
13
|
csv_sha1="01c312c4c16f7acc5afddaafcf826e5af58b13e2",
|
14
14
|
zip_sha1="1c070b88fa45e9e69d58f95466cb6406a45a4873",
|
15
|
-
csv_version_id="
|
16
|
-
zip_version_id="
|
15
|
+
csv_version_id="null",
|
16
|
+
zip_version_id="null")
|
17
17
|
|
18
18
|
# fitting stimuli
|
19
19
|
stimulus_set_registry['Coggan2024_behavior_fitting'] = lambda: (
|
20
20
|
load_stimulus_set_from_s3(
|
21
21
|
identifier="tong.Coggan2024_behavior_fitting",
|
22
|
-
bucket="brainio-brainscore",
|
22
|
+
bucket="brainscore-storage/brainio-brainscore",
|
23
23
|
csv_sha1="136e48992305ea78a4fb77e9dfc75dcf01e885d0",
|
24
24
|
zip_sha1="24e68f5ba2f8f2105daf706307642637118e7d36",
|
25
|
-
csv_version_id="
|
26
|
-
zip_version_id="
|
25
|
+
csv_version_id="null",
|
26
|
+
zip_version_id="null"))
|
27
27
|
|
28
28
|
# behavioral data
|
29
29
|
data_registry['Coggan2024_behavior'] = lambda: load_assembly_from_s3(
|
30
30
|
identifier="tong.Coggan2024_behavior",
|
31
|
-
version_id="
|
31
|
+
version_id="null",
|
32
32
|
sha1="c1ac4a268476c35bbe40081358667a03d3544631",
|
33
|
-
bucket="brainio-brainscore",
|
33
|
+
bucket="brainscore-storage/brainio-brainscore",
|
34
34
|
cls=BehavioralAssembly,
|
35
35
|
stimulus_set_loader=lambda: load_stimulus_set('Coggan2024_behavior'),
|
36
36
|
)
|
@@ -138,7 +138,7 @@ packaged_stimulus_metadata = package_stimulus_set(
|
|
138
138
|
catalog_name=None,
|
139
139
|
proto_stimulus_set=stimuli,
|
140
140
|
stimulus_set_identifier=stimuli.name,
|
141
|
-
bucket_name="brainio-brainscore")
|
141
|
+
bucket_name="brainscore-storage/brainio-brainscore")
|
142
142
|
print(packaged_stimulus_metadata)
|
143
143
|
|
144
144
|
|
@@ -161,6 +161,6 @@ packaged_behavioral_metadata = package_data_assembly(
|
|
161
161
|
assembly_identifier=assembly.name,
|
162
162
|
stimulus_set_identifier=stimuli.name,
|
163
163
|
assembly_class_name="BehavioralAssembly",
|
164
|
-
bucket_name="brainio-brainscore",
|
164
|
+
bucket_name="brainscore-storage/brainio-brainscore",
|
165
165
|
catalog_identifier=None)
|
166
166
|
print(packaged_behavioral_metadata)
|
@@ -9,19 +9,18 @@ from brainscore_vision.data_helpers.s3 import (
|
|
9
9
|
# stimulus set
|
10
10
|
stimulus_set_registry['Coggan2024_fMRI'] = lambda: load_stimulus_set_from_s3(
|
11
11
|
identifier="tong.Coggan2024_fMRI",
|
12
|
-
bucket="brainio-brainscore",
|
12
|
+
bucket="brainscore-storage/brainio-brainscore",
|
13
13
|
csv_sha1="0089f5f8fd3f2de14de12ed736a0f88575f8e1ee",
|
14
14
|
zip_sha1="e26fdea4d866799526dea183f5bfb9792718822a",
|
15
|
-
csv_version_id="
|
16
|
-
zip_version_id="
|
15
|
+
csv_version_id="kAS9zqRPf7zmLCN73I2VB5gBvhrZoU0p",
|
16
|
+
zip_version_id="jmiuKdlsh6.1suR_5FvLFDyg.PCzEZTn")
|
17
17
|
|
18
18
|
# fMRI data
|
19
19
|
data_registry['Coggan2024_fMRI'] = lambda: load_assembly_from_s3(
|
20
20
|
identifier="tong.Coggan2024_fMRI",
|
21
|
-
version_id="
|
21
|
+
version_id="LRYpTHYG7IoNCexbzZdQM4_LjML_M3CS",
|
22
22
|
sha1="da3adbca5247d0491d366f94e8431fb3e4e58db2",
|
23
|
-
bucket="brainio-brainscore",
|
23
|
+
bucket="brainscore-storage/brainio-brainscore",
|
24
24
|
cls=NeuroidAssembly,
|
25
25
|
stimulus_set_loader=lambda: load_stimulus_set('Coggan2024_fMRI'),
|
26
26
|
)
|
27
|
-
|
@@ -81,7 +81,7 @@ packaged_stimulus_metadata = package_stimulus_set(
|
|
81
81
|
catalog_name=None,
|
82
82
|
proto_stimulus_set=stimuli,
|
83
83
|
stimulus_set_identifier=stimuli.name,
|
84
|
-
bucket_name="brainio-brainscore")
|
84
|
+
bucket_name="brainscore-storage/brainio-brainscore")
|
85
85
|
pkl.dump(packaged_stimulus_metadata, open('packaged_stimulus_metadata.pkl', 'wb'))
|
86
86
|
print(packaged_stimulus_metadata)
|
87
87
|
"""
|
@@ -115,7 +115,7 @@ packaged_neural_metadata = package_data_assembly(
|
|
115
115
|
assembly_identifier=assembly.name,
|
116
116
|
stimulus_set_identifier=stimuli.name,
|
117
117
|
assembly_class_name="NeuroidAssembly",
|
118
|
-
bucket_name="brainio-brainscore",
|
118
|
+
bucket_name="brainscore-storage/brainio-brainscore",
|
119
119
|
catalog_identifier=None)
|
120
120
|
|
121
121
|
# save the packaged metadata
|
@@ -17,9 +17,9 @@ BIBTEX = """@article{david2004evaluation,
|
|
17
17
|
# assembly
|
18
18
|
data_registry['David2004'] = lambda: load_assembly_from_s3(
|
19
19
|
identifier="gallant.David2004",
|
20
|
-
version_id="
|
20
|
+
version_id="null",
|
21
21
|
sha1="d2ed9834c054da2333f5d894285c9841a1f27313",
|
22
|
-
bucket="brainio-brainscore",
|
22
|
+
bucket="brainscore-storage/brainio-brainscore",
|
23
23
|
cls=NeuronRecordingAssembly,
|
24
24
|
stimulus_set_loader=lambda: load_stimulus_set('David2004'),
|
25
25
|
)
|
@@ -27,8 +27,8 @@ data_registry['David2004'] = lambda: load_assembly_from_s3(
|
|
27
27
|
# stimulus set
|
28
28
|
stimulus_set_registry['David2004'] = lambda: load_stimulus_set_from_s3(
|
29
29
|
identifier="David2004",
|
30
|
-
bucket="brainio-brainscore",
|
30
|
+
bucket="brainscore-storage/brainio-brainscore",
|
31
31
|
csv_sha1="8ec76338b998cadcdf1e57edd2dd992e2ab2355b",
|
32
32
|
zip_sha1="0200421d66a0613946d39cab64c00b561160016e",
|
33
|
-
csv_version_id="
|
34
|
-
zip_version_id="
|
33
|
+
csv_version_id="null",
|
34
|
+
zip_version_id="null")
|
@@ -15,8 +15,8 @@ BIBTEX = """@INPROCEEDINGS{5206848,
|
|
15
15
|
# stimulus set
|
16
16
|
stimulus_set_registry['imagenet_val'] = lambda: load_stimulus_set_from_s3(
|
17
17
|
identifier="imagenet_val",
|
18
|
-
bucket="brainio-brainscore",
|
18
|
+
bucket="brainscore-storage/brainio-brainscore",
|
19
19
|
csv_sha1="ff79dcf6b0d115e6e8aa8d0fbba3af11dc649e57",
|
20
20
|
zip_sha1="78172d752d8216a00833cfa34be67c8532ad7330",
|
21
|
-
csv_version_id="
|
22
|
-
zip_version_id="
|
21
|
+
csv_version_id="null",
|
22
|
+
zip_version_id="null")
|