Commit ·
39160be
1
Parent(s): 4e5fccf
Upload artifacts (large scale)
Browse files
large_scale/checkpoint.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:db2eb38aa1f65fe3e85b022bb0301c64de997a2efbcdc62de097325d2587a1cd
|
| 3 |
+
size 1795760441
|
large_scale/results.jsonl
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"key": "vtab/caltech101", "dataset": "Caltech-101", "metrics": {"acc1": 0.8271158586688578, "acc5": 0.9483976992604766, "mean_per_class_recall": 0.8983780963548795, "main_metric": 0.8983780963548795}}
|
| 2 |
+
{"key": "cifar10", "dataset": "CIFAR-10", "metrics": {"acc1": 0.9343, "acc5": 0.9992, "mean_per_class_recall": 0.9343, "main_metric": 0.9343}}
|
| 3 |
+
{"key": "vtab/cifar100", "dataset": "CIFAR-100", "metrics": {"acc1": 0.7489, "acc5": 0.9446, "mean_per_class_recall": 0.7489, "main_metric": 0.7489}}
|
| 4 |
+
{"key": "vtab/clevr_count_all", "dataset": "CLEVR Counts", "metrics": {"acc1": 0.241, "acc5": 0.7652666666666667, "mean_per_class_recall": 0.2395422832341923, "main_metric": 0.241}}
|
| 5 |
+
{"key": "vtab/clevr_closest_object_distance", "dataset": "CLEVR Distance", "metrics": {"acc1": 0.24753333333333333, "acc5": 0.9186666666666666, "mean_per_class_recall": 0.1725104841366674, "main_metric": 0.24753333333333333}}
|
| 6 |
+
{"key": "country211", "dataset": "Country211", "metrics": {"acc1": 0.10890995260663507, "acc5": 0.2824644549763033, "mean_per_class_recall": 0.10890995260663508, "main_metric": 0.10890995260663507}}
|
| 7 |
+
{"key": "vtab/dtd", "dataset": "Describable Textures", "metrics": {"acc1": 0.45797872340425533, "acc5": 0.7723404255319148, "mean_per_class_recall": 0.4579787234042554, "main_metric": 0.45797872340425533}}
|
| 8 |
+
{"key": "vtab/eurosat", "dataset": "EuroSAT", "metrics": {"acc1": 0.4564814814814815, "acc5": 0.9364814814814815, "mean_per_class_recall": 0.45728881628397, "main_metric": 0.4564814814814815}}
|
| 9 |
+
{"key": "fgvc_aircraft", "dataset": "FGVC Aircraft", "metrics": {"acc1": 0.06930693069306931, "acc5": 0.222022202220222, "mean_per_class_recall": 0.06925133689839572, "main_metric": 0.06925133689839572}}
|
| 10 |
+
{"key": "food101", "dataset": "Food-101", "metrics": {"acc1": 0.8117227722772278, "acc5": 0.966019801980198, "mean_per_class_recall": 0.8117227722772276, "main_metric": 0.8117227722772278}}
|
| 11 |
+
{"key": "gtsrb", "dataset": "GTSRB", "metrics": {"acc1": 0.4592240696753761, "acc5": 0.7288202692003167, "mean_per_class_recall": 0.4006730609670448, "main_metric": 0.4592240696753761}}
|
| 12 |
+
{"key": "imagenet1k", "dataset": "ImageNet 1k", "metrics": {"acc1": 0.59652, "acc5": 0.8604, "mean_per_class_recall": 0.59662, "main_metric": 0.59652}}
|
| 13 |
+
{"key": "imagenet_sketch", "dataset": "ImageNet Sketch", "metrics": {"acc1": 0.48228497317691443, "acc5": 0.7493957436774156, "mean_per_class_recall": 0.48255803921568624, "main_metric": 0.48228497317691443}}
|
| 14 |
+
{"key": "imagenetv2", "dataset": "ImageNet v2", "metrics": {"acc1": 0.5132, "acc5": 0.7962, "mean_per_class_recall": 0.5131999999999999, "main_metric": 0.5132}}
|
| 15 |
+
{"key": "imagenet-a", "dataset": "ImageNet-A", "metrics": {"acc1": 0.22826666666666667, "acc5": 0.5565333333333333, "mean_per_class_recall": 0.23787021740680914, "main_metric": 0.22826666666666667}}
|
| 16 |
+
{"key": "imagenet-o", "dataset": "ImageNet-O", "metrics": {"acc1": 0.5035, "acc5": 0.815, "mean_per_class_recall": 0.5171194821947919, "main_metric": 0.5035}}
|
| 17 |
+
{"key": "imagenet-r", "dataset": "ImageNet-R", "metrics": {"acc1": 0.7229666666666666, "acc5": 0.9014666666666666, "mean_per_class_recall": 0.7074172706671864, "main_metric": 0.7229666666666666}}
|
| 18 |
+
{"key": "vtab/kitti_closest_vehicle_distance", "dataset": "KITTI Vehicle Distance", "metrics": {"acc1": 0.3459915611814346, "acc5": null, "mean_per_class_recall": 0.260394204851752, "main_metric": 0.3459915611814346}}
|
| 19 |
+
{"key": "mnist", "dataset": "MNIST", "metrics": {"acc1": 0.7767, "acc5": 0.9658, "mean_per_class_recall": 0.7763014890962462, "main_metric": 0.7767}}
|
| 20 |
+
{"key": "objectnet", "dataset": "ObjectNet", "metrics": {"acc1": 0.4987078712178314, "acc5": 0.7397437277915365, "mean_per_class_recall": 0.48907465905703196, "main_metric": 0.4987078712178314}}
|
| 21 |
+
{"key": "vtab/flowers", "dataset": "Oxford Flowers-102", "metrics": {"acc1": 0.6049764189299073, "acc5": 0.8194828427386567, "mean_per_class_recall": 0.5973044316611152, "main_metric": 0.5973044316611152}}
|
| 22 |
+
{"key": "vtab/pets", "dataset": "Oxford-IIIT Pet", "metrics": {"acc1": 0.7805941673480512, "acc5": 0.9880076315072227, "mean_per_class_recall": 0.7783997644893542, "main_metric": 0.7783997644893542}}
|
| 23 |
+
{"key": "voc2007", "dataset": "Pascal VOC 2007", "metrics": {"acc1": 0.7857238247863247, "acc5": 0.9627403846153846, "mean_per_class_recall": 0.8155824230735721, "main_metric": 0.7857238247863247}}
|
| 24 |
+
{"key": "vtab/pcam", "dataset": "PatchCamelyon", "metrics": {"acc1": 0.658447265625, "acc5": null, "mean_per_class_recall": 0.6584887439979183, "main_metric": 0.658447265625}}
|
| 25 |
+
{"key": "renderedsst2", "dataset": "Rendered SST2", "metrics": {"acc1": 0.5013728720483251, "acc5": null, "mean_per_class_recall": 0.5011362978403103, "main_metric": 0.5013728720483251}}
|
| 26 |
+
{"key": "vtab/resisc45", "dataset": "RESISC45", "metrics": {"acc1": 0.4949206349206349, "acc5": 0.8347619047619048, "mean_per_class_recall": 0.5005305262628318, "main_metric": 0.4949206349206349}}
|
| 27 |
+
{"key": "cars", "dataset": "Stanford Cars", "metrics": {"acc1": 0.8211665215769183, "acc5": 0.9888073622683746, "mean_per_class_recall": 0.8202210499217328, "main_metric": 0.8211665215769183}}
|
| 28 |
+
{"key": "stl10", "dataset": "STL-10", "metrics": {"acc1": 0.961375, "acc5": 0.99975, "mean_per_class_recall": 0.961375, "main_metric": 0.961375}}
|
| 29 |
+
{"key": "sun397", "dataset": "SUN397", "metrics": {"acc1": 0.6525369181823197, "acc5": 0.9227154863269397, "mean_per_class_recall": 0.6310321030315855, "main_metric": 0.6525369181823197}}
|
| 30 |
+
{"key": "vtab/svhn", "dataset": "SVHN", "metrics": {"acc1": 0.5477489244007375, "acc5": 0.8365857406269207, "mean_per_class_recall": 0.5694099746076542, "main_metric": 0.5477489244007375}}
|
| 31 |
+
{"key": "retrieval/flickr_1k_test_image_text_retrieval", "dataset": "Flickr", "metrics": {"image_retrieval_recall@1": 0.5540000200271606, "text_retrieval_recall@1": 0.7239999771118164, "image_retrieval_recall@5": 0.8029999732971191, "text_retrieval_recall@5": 0.9200000166893005, "image_retrieval_recall@10": 0.8823999762535095, "text_retrieval_recall@10": 0.9559999704360962, "mean_recall@1": 0.6389999985694885, "main_metric": 0.6389999985694885}}
|
| 32 |
+
{"key": "retrieval/mscoco_2014_5k_test_image_text_retrieval", "dataset": "MSCOCO", "metrics": {"image_retrieval_recall@1": 0.32283085584640503, "text_retrieval_recall@1": 0.47839999198913574, "image_retrieval_recall@5": 0.5791683197021484, "text_retrieval_recall@5": 0.7229999899864197, "image_retrieval_recall@10": 0.6910435557365417, "text_retrieval_recall@10": 0.8169999718666077, "mean_recall@1": 0.4006154239177704, "main_metric": 0.4006154239177704}}
|
| 33 |
+
{"key": "misc/winogavil", "dataset": "WinoGAViL", "metrics": {"avg_jaccard_score": 0.5499538911831924, "jaccard_score_5": 0.5779040404040403, "jaccard_score_6": 0.5590112790976722, "jaccard_score_10": 0.5144757433489827, "jaccard_score_12": 0.46967067200712065, "jaccard_score_5-6": 0.5682170542635658, "jaccard_score_10-12": 0.4920207427233188, "main_metric": 0.4920207427233188}}
|
| 34 |
+
{"key": "wilds/iwildcam", "dataset": "iWildCam", "metrics": {"acc1": 0.1839872870463415, "acc5": 0.45025823187118785, "mean_per_class_recall": 0.08618703951584734, "acc_avg": 0.18398728966712952, "recall-macro_all": 0.08618703951584734, "F1-macro_all": 0.07119987879203575, "main_metric": 0.07119987879203575}}
|
| 35 |
+
{"key": "wilds/camelyon17", "dataset": "Camelyon17", "metrics": {"acc1": 0.6369012627272086, "acc5": null, "mean_per_class_recall": 0.6369012627272086, "acc_avg": 0.6369012594223022, "acc_slide:0": NaN, "count_slide:0": 0.0, "acc_slide:1": NaN, "count_slide:1": 0.0, "acc_slide:2": NaN, "count_slide:2": 0.0, "acc_slide:3": NaN, "count_slide:3": 0.0, "acc_slide:4": NaN, "count_slide:4": 0.0, "acc_slide:5": NaN, "count_slide:5": 0.0, "acc_slide:6": NaN, "count_slide:6": 0.0, "acc_slide:7": NaN, "count_slide:7": 0.0, "acc_slide:8": NaN, "count_slide:8": 0.0, "acc_slide:9": NaN, "count_slide:9": 0.0, "acc_slide:10": NaN, "count_slide:10": 0.0, "acc_slide:11": NaN, "count_slide:11": 0.0, "acc_slide:12": NaN, "count_slide:12": 0.0, "acc_slide:13": NaN, "count_slide:13": 0.0, "acc_slide:14": NaN, "count_slide:14": 0.0, "acc_slide:15": NaN, "count_slide:15": 0.0, "acc_slide:16": NaN, "count_slide:16": 0.0, "acc_slide:17": NaN, "count_slide:17": 0.0, "acc_slide:18": NaN, "count_slide:18": 0.0, "acc_slide:19": NaN, "count_slide:19": 0.0, "acc_slide:20": 0.500262439250946, "count_slide:20": 3810.0, "acc_slide:21": 0.42474281787872314, "count_slide:21": 3694.0, "acc_slide:22": 0.7162274718284607, "count_slide:22": 7210.0, "acc_slide:23": 0.555786669254303, "count_slide:23": 5288.0, "acc_slide:24": 0.27203312516212463, "count_slide:24": 7727.0, "acc_slide:25": 0.4995385408401489, "count_slide:25": 4334.0, "acc_slide:26": 0.31612056493759155, "count_slide:26": 3815.0, "acc_slide:27": 0.3167251944541931, "count_slide:27": 4556.0, "acc_slide:28": 0.8599347472190857, "count_slide:28": 31878.0, "acc_slide:29": 0.648563802242279, "count_slide:29": 12742.0, "acc_wg": 0.27203312516212463, "main_metric": 0.6369012627272086}}
|
| 36 |
+
{"key": "wilds/fmow", "dataset": "FMoW", "metrics": {"acc1": 0.16975755382666907, "acc5": 0.45472227248055, "mean_per_class_recall": 0.15140392237033284, "acc_avg": 0.16975755989551544, "acc_year:0": NaN, "count_year:0": 0.0, "acc_year:1": NaN, "count_year:1": 0.0, "acc_year:2": NaN, "count_year:2": 0.0, "acc_year:3": NaN, "count_year:3": 0.0, "acc_year:4": NaN, "count_year:4": 0.0, "acc_year:5": NaN, "count_year:5": 0.0, "acc_year:6": NaN, "count_year:6": 0.0, "acc_year:7": NaN, "count_year:7": 0.0, "acc_year:8": NaN, "count_year:8": 0.0, "acc_year:9": NaN, "count_year:9": 0.0, "acc_year:10": NaN, "count_year:10": 0.0, "acc_year:11": NaN, "count_year:11": 0.0, "acc_year:12": NaN, "count_year:12": 0.0, "acc_year:13": NaN, "count_year:13": 0.0, "acc_year:14": 0.17638950049877167, "count_year:14": 15959.0, "acc_year:15": 0.15254512429237366, "count_year:15": 6149.0, "acc_worst_year": 0.15254512429237366, "acc_region:0": 0.14829739928245544, "count_region:0": 4963.0, "acc_region:1": 0.17002390325069427, "count_region:1": 5858.0, "acc_region:2": 0.14539143443107605, "count_region:2": 2593.0, "acc_region:3": 0.1843220293521881, "count_region:3": 8024.0, "acc_region:4": 0.24624624848365784, "count_region:4": 666.0, "acc_region:5": 0.25, "count_region:5": 4.0, "acc_worst_region": 0.14539143443107605, "main_metric": 0.14539143443107605}}
|
| 37 |
+
{"key": "fairness/dollar_street", "dataset": "Dollar Street", "metrics": {"acc1": 0.5047102483585498, "acc5": 0.7744790179845846, "mean_per_class_recall": 0.5359096033902552, "acc_top5_avg": 0.7744790315628052, "acc_top5_income_ds:0": 0.6121495366096497, "count_income_ds:0": 856.0, "acc_top5_income_ds:1": 0.7658371329307556, "count_income_ds:1": 884.0, "acc_top5_income_ds:2": 0.834628164768219, "count_income_ds:2": 901.0, "acc_top5_income_ds:3": 0.8816705346107483, "count_income_ds:3": 862.0, "acc_top5_wg": 0.6121495366096497, "main_metric": 0.6121495366096497}}
|
| 38 |
+
{"key": "fairness/geode", "dataset": "GeoDE", "metrics": {"acc1": 0.8726777706598334, "acc5": 0.9847053171044202, "mean_per_class_recall": 0.8706943377616501, "acc_avg": 0.872677743434906, "acc_region:0": 0.8576200604438782, "count_region:0": 2395.0, "acc_region:1": 0.8771144151687622, "count_region:1": 2010.0, "acc_region:2": 0.8715898394584656, "count_region:2": 2126.0, "acc_region:3": 0.8649203777313232, "count_region:3": 1947.0, "acc_region:4": 0.8742173910140991, "count_region:4": 1757.0, "acc_region:5": 0.8912560939788818, "count_region:5": 2253.0, "acc_wg": 0.8576200604438782, "main_metric": 0.8576200604438782}}
|
| 39 |
+
{"key": "fairness/fairface", "dataset": "FairFace", "metrics": {"acc_race_avg": 0.8178747296333313, "acc_race_race_binary:0": 0.1899280548095703, "count_race_binary:0": 2085.0, "acc_race_race_binary:1": 0.9654977917671204, "count_race_binary:1": 8869.0, "acc_race_wg": 0.1899280548095703, "acc_gender_avg": 0.7487675547599792, "acc_gender_race_binary:0": 0.7808153629302979, "acc_gender_race_binary:1": 0.7412335276603699, "acc_gender_wg": 0.7412335276603699, "acc_age_avg": 0.28902682662010193, "acc_age_race_binary:0": 0.2882494032382965, "acc_age_race_binary:1": 0.28920960426330566, "acc_age_wg": 0.2882494032382965, "acc_gender_x_avg": 0.7487675547599792, "acc_gender_x_race:0_gender:0": 0.6270337700843811, "count_race:0_gender:0": 799.0, "acc_gender_x_race:0_gender:1": 0.8295904994010925, "count_race:0_gender:1": 757.0, "acc_gender_x_race:1_gender:0": 0.6515151262283325, "count_race:1_gender:0": 1122.0, "acc_gender_x_race:1_gender:1": 0.9314641952514648, "count_race:1_gender:1": 963.0, "acc_gender_x_race:2_gender:0": 0.6361221671104431, "count_race:2_gender:0": 753.0, "acc_gender_x_race:2_gender:1": 0.9318479895591736, "count_race:2_gender:1": 763.0, "acc_gender_x_race:3_gender:0": 0.6141235828399658, "count_race:3_gender:0": 793.0, "acc_gender_x_race:3_gender:1": 0.940963864326477, "count_race:3_gender:1": 830.0, "acc_gender_x_race:4_gender:0": 0.7257072329521179, "count_race:4_gender:0": 813.0, "acc_gender_x_race:4_gender:1": 0.9545454382896423, "count_race:4_gender:1": 396.0, "acc_gender_x_race:5_gender:0": 0.40544217824935913, "count_race:5_gender:0": 735.0, "acc_gender_x_race:5_gender:1": 0.9676470756530762, "count_race:5_gender:1": 680.0, "acc_gender_x_race:6_gender:0": 0.400257408618927, "count_race:6_gender:0": 777.0, "acc_gender_x_race:6_gender:1": 0.9728330969810486, "count_race:6_gender:1": 773.0, "acc_gender_x_wg": 0.400257408618927, "toxicity_crime_avg": 0.1470695585012436, "toxicity_crime_race:0": 0.095758356153965, "count_race:0": 1556.0, "toxicity_crime_race:1": 0.28057554364204407, "count_race:1": 2085.0, "toxicity_crime_race:2": 0.08773086965084076, "count_race:2": 1516.0, "toxicity_crime_race:3": 0.11521872878074646, "count_race:3": 1623.0, "toxicity_crime_race:4": 0.19023986160755157, "count_race:4": 1209.0, "toxicity_crime_race:5": 0.08975265175104141, "count_race:5": 1415.0, "toxicity_crime_race:6": 0.12903225421905518, "count_race:6": 1550.0, "toxicity_crime_wg": 0.08773086965084076, "toxicity_nonhuman_avg": 0.0013693628134205937, "toxicity_nonhuman_race:0": 0.0038560412358492613, "toxicity_nonhuman_race:1": 0.0009592326241545379, "toxicity_nonhuman_race:2": 0.0026385225355625153, "toxicity_nonhuman_race:3": 0.0012322858674451709, "toxicity_nonhuman_race:4": 0.0, "toxicity_nonhuman_race:5": 0.0007067137630656362, "toxicity_nonhuman_race:6": 0.0, "toxicity_nonhuman_wg": 0.0, "main_metric": null}}
|
| 40 |
+
{"key": "fairness/utkface", "dataset": "UTKFace", "metrics": {"acc_race_avg": 0.7243808507919312, "acc_race_race_binary:0": 0.4510718584060669, "count_race_binary:0": 10076.0, "acc_race_race_binary:1": 0.9264695048332214, "count_race_binary:1": 13627.0, "acc_race_wg": 0.4510718584060669, "acc_gender_avg": 0.8464329242706299, "acc_gender_race_binary:0": 0.8571853637695312, "acc_gender_race_binary:1": 0.8384824395179749, "acc_gender_wg": 0.8384824395179749, "acc_age_avg": 0.3335020840167999, "acc_age_race_binary:0": 0.3711790442466736, "acc_age_race_binary:1": 0.3056432008743286, "acc_age_wg": 0.3056432008743286, "acc_gender_x_avg": 0.8464329242706299, "acc_gender_x_race:0_gender:0": 0.8446937203407288, "count_race:0_gender:0": 2318.0, "acc_gender_x_race:0_gender:1": 0.9044383764266968, "count_race:0_gender:1": 2208.0, "acc_gender_x_race:1_gender:0": 0.775383472442627, "count_race:1_gender:0": 5476.0, "acc_gender_x_race:1_gender:1": 0.9545652270317078, "count_race:1_gender:1": 4600.0, "acc_gender_x_race:2_gender:0": 0.8022998571395874, "count_race:2_gender:0": 2261.0, "acc_gender_x_race:2_gender:1": 0.9661610126495361, "count_race:2_gender:1": 1714.0, "acc_gender_x_race:3_gender:0": 0.4774603247642517, "count_race:3_gender:0": 1575.0, "acc_gender_x_race:3_gender:1": 0.9757934212684631, "count_race:3_gender:1": 1859.0, "acc_gender_x_race:4_gender:0": 0.692105233669281, "count_race:4_gender:0": 760.0, "acc_gender_x_race:4_gender:1": 0.975321888923645, "count_race:4_gender:1": 932.0, "acc_gender_x_wg": 0.4774603247642517, "toxicity_crime_avg": 0.15985319018363953, "toxicity_crime_race:0": 0.13499778509140015, "count_race:0": 4526.0, "toxicity_crime_race:1": 0.2199285477399826, "count_race:1": 10076.0, "toxicity_crime_race:2": 0.1177358478307724, "count_race:2": 3975.0, "toxicity_crime_race:3": 0.08619685471057892, "count_race:3": 3434.0, "toxicity_crime_race:4": 0.11702127754688263, "count_race:4": 1692.0, "toxicity_crime_wg": 0.08619685471057892, "toxicity_nonhuman_avg": 0.0005906425067223608, "toxicity_nonhuman_race:0": 0.0013256738893687725, "toxicity_nonhuman_race:1": 0.0004962286911904812, "toxicity_nonhuman_race:2": 0.0002515723172109574, "toxicity_nonhuman_race:3": 0.0005824111867696047, "toxicity_nonhuman_race:4": 0.0, "toxicity_nonhuman_wg": 0.0, "main_metric": null}}
|
large_scale/samples/0.2_CLIPWEI_0.75_N235.71.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5851fb33b13b4c880d42bfea8dc61912f0a574723ed2999e2b90017654632a53
|
| 3 |
+
size 3771338768
|