"...keras-cv/keras_cv/layers/regularization/squeeze_excite.py" did not exist on "9846958aa60286f18e9741dc2b0c1e325aaa832a"
mobileclip_s2.jsonl 16.1 KB
Newer Older
chenzk's avatar
v1.0  
chenzk committed
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
{"key": "vtab/caltech101", "dataset": "Caltech-101", "metrics": {"acc1": 0.8447000821692687, "acc5": 0.9541495480690222, "mean_per_class_recall": 0.936722834577359, "main_metric": 0.936722834577359}}
{"key": "cifar10", "dataset": "CIFAR-10", "metrics": {"acc1": 0.9588, "acc5": 0.9995, "mean_per_class_recall": 0.9588000000000001, "main_metric": 0.9588}}
{"key": "vtab/cifar100", "dataset": "CIFAR-100", "metrics": {"acc1": 0.8128, "acc5": 0.9667, "mean_per_class_recall": 0.8128, "main_metric": 0.8128}}
{"key": "vtab/clevr_count_all", "dataset": "CLEVR Counts", "metrics": {"acc1": 0.3344, "acc5": 0.9235333333333333, "mean_per_class_recall": 0.33309194769139167, "main_metric": 0.3344}}
{"key": "vtab/clevr_closest_object_distance", "dataset": "CLEVR Distance", "metrics": {"acc1": 0.2376, "acc5": 0.9186666666666666, "mean_per_class_recall": 0.17236004590330653, "main_metric": 0.2376}}
{"key": "country211", "dataset": "Country211", "metrics": {"acc1": 0.19199052132701422, "acc5": 0.42241706161137443, "mean_per_class_recall": 0.19199052132701416, "main_metric": 0.19199052132701422}}
{"key": "vtab/dtd", "dataset": "Describable Textures", "metrics": {"acc1": 0.6090425531914894, "acc5": 0.8803191489361702, "mean_per_class_recall": 0.6090425531914894, "main_metric": 0.6090425531914894}}
{"key": "vtab/eurosat", "dataset": "EuroSAT", "metrics": {"acc1": 0.6672222222222223, "acc5": 0.9757407407407407, "mean_per_class_recall": 0.6710678146515698, "main_metric": 0.6672222222222223}}
{"key": "fgvc_aircraft", "dataset": "FGVC Aircraft", "metrics": {"acc1": 0.30423042304230424, "acc5": 0.6753675367536753, "mean_per_class_recall": 0.30353832442067735, "main_metric": 0.30353832442067735}}
{"key": "food101", "dataset": "Food-101", "metrics": {"acc1": 0.8967920792079208, "acc5": 0.9872475247524752, "mean_per_class_recall": 0.8967920792079209, "main_metric": 0.8967920792079208}}
{"key": "gtsrb", "dataset": "GTSRB", "metrics": {"acc1": 0.5361836896278701, "acc5": 0.8277117973079968, "mean_per_class_recall": 0.4787885812648198, "main_metric": 0.5361836896278701}}
{"key": "imagenet1k", "dataset": "ImageNet 1k", "metrics": {"acc1": 0.74438, "acc5": 0.94396, "mean_per_class_recall": 0.74424, "main_metric": 0.74438}}
{"key": "imagenet_sketch", "dataset": "ImageNet Sketch", "metrics": {"acc1": 0.6221187289984083, "acc5": 0.8650985478197646, "mean_per_class_recall": 0.6223282352941176, "main_metric": 0.6221187289984083}}
{"key": "imagenetv2", "dataset": "ImageNet v2", "metrics": {"acc1": 0.6679, "acc5": 0.9042, "mean_per_class_recall": 0.6684000000000001, "main_metric": 0.6679}}
{"key": "imagenet-a", "dataset": "ImageNet-A", "metrics": {"acc1": 0.4932, "acc5": 0.8016, "mean_per_class_recall": 0.4960668896772007, "main_metric": 0.4932}}
{"key": "imagenet-o", "dataset": "ImageNet-O", "metrics": {"acc1": 0.469, "acc5": 0.7885, "mean_per_class_recall": 0.48375934665609593, "main_metric": 0.469}}
{"key": "imagenet-r", "dataset": "ImageNet-R", "metrics": {"acc1": 0.8698666666666667, "acc5": 0.9681333333333333, "mean_per_class_recall": 0.8553672907748989, "main_metric": 0.8698666666666667}}
{"key": "vtab/kitti_closest_vehicle_distance", "dataset": "KITTI Vehicle Distance", "metrics": {"acc1": 0.2320675105485232, "acc5": null, "mean_per_class_recall": 0.32955606784304503, "main_metric": 0.2320675105485232}}
{"key": "mnist", "dataset": "MNIST", "metrics": {"acc1": 0.8702, "acc5": 0.9823, "mean_per_class_recall": 0.8687912186633511, "main_metric": 0.8702}}
{"key": "objectnet", "dataset": "ObjectNet", "metrics": {"acc1": 0.6656078389146118, "acc5": 0.8666953806396037, "mean_per_class_recall": 0.6523596249996781, "main_metric": 0.6656078389146118}}
{"key": "vtab/flowers", "dataset": "Oxford Flowers-102", "metrics": {"acc1": 0.7503659131566108, "acc5": 0.88209464953651, "mean_per_class_recall": 0.7320917902269698, "main_metric": 0.7320917902269698}}
{"key": "vtab/pets", "dataset": "Oxford-IIIT Pet", "metrics": {"acc1": 0.9283183428727174, "acc5": 0.9983646770237122, "mean_per_class_recall": 0.9278175397989731, "main_metric": 0.9278175397989731}}
{"key": "voc2007", "dataset": "Pascal VOC 2007", "metrics": {"acc1": 0.8321981837606838, "acc5": 0.9822382478632479, "mean_per_class_recall": 0.8719232098874825, "main_metric": 0.8321981837606838}}
{"key": "vtab/pcam", "dataset": "PatchCamelyon", "metrics": {"acc1": 0.68829345703125, "acc5": null, "mean_per_class_recall": 0.6882425499107128, "main_metric": 0.68829345703125}}
{"key": "renderedsst2", "dataset": "Rendered SST2", "metrics": {"acc1": 0.5634266886326195, "acc5": null, "mean_per_class_recall": 0.5631477621446355, "main_metric": 0.5634266886326195}}
{"key": "vtab/resisc45", "dataset": "RESISC45", "metrics": {"acc1": 0.686984126984127, "acc5": 0.9433333333333334, "mean_per_class_recall": 0.6934871572095898, "main_metric": 0.686984126984127}}
{"key": "cars", "dataset": "Stanford Cars", "metrics": {"acc1": 0.8612112921278448, "acc5": 0.9953985822658873, "mean_per_class_recall": 0.8615237612708011, "main_metric": 0.8612112921278448}}
{"key": "stl10", "dataset": "STL-10", "metrics": {"acc1": 0.987375, "acc5": 1.0, "mean_per_class_recall": 0.9873749999999999, "main_metric": 0.987375}}
{"key": "sun397", "dataset": "SUN397", "metrics": {"acc1": 0.7306857678798021, "acc5": 0.9532063188480424, "mean_per_class_recall": 0.7194289168008176, "main_metric": 0.7306857678798021}}
{"key": "vtab/svhn", "dataset": "SVHN", "metrics": {"acc1": 0.68899815611555, "acc5": 0.9524815611555009, "mean_per_class_recall": 0.6898340759970616, "main_metric": 0.68899815611555}}
{"key": "retrieval/flickr_1k_test_image_text_retrieval", "dataset": "Flickr", "metrics": {"image_retrieval_recall@1": 0.7337999939918518, "text_retrieval_recall@1": 0.902999997138977, "image_retrieval_recall@5": 0.9229999780654907, "text_retrieval_recall@5": 0.9890000224113464, "image_retrieval_recall@10": 0.9562000036239624, "text_retrieval_recall@10": 0.9959999918937683, "mean_recall@1": 0.8183999955654144, "main_metric": 0.8183999955654144}}
{"key": "retrieval/mscoco_2014_5k_test_image_text_retrieval", "dataset": "MSCOCO", "metrics": {"image_retrieval_recall@1": 0.4537385106086731, "text_retrieval_recall@1": 0.6335999965667725, "image_retrieval_recall@5": 0.7005198001861572, "text_retrieval_recall@5": 0.8514000177383423, "image_retrieval_recall@10": 0.7901239395141602, "text_retrieval_recall@10": 0.9138000011444092, "mean_recall@1": 0.5436692535877228, "main_metric": 0.5436692535877228}}
{"key": "misc/winogavil", "dataset": "WinoGAViL", "metrics": {"avg_jaccard_score": 0.546845221389145, "jaccard_score_5": 0.5773989898989897, "jaccard_score_6": 0.561987041036717, "jaccard_score_10": 0.48857589984350547, "jaccard_score_12": 0.4614708500222519, "jaccard_score_5-6": 0.569496739264181, "jaccard_score_10-12": 0.47499163599866173, "main_metric": 0.47499163599866173}}
{"key": "wilds/iwildcam", "dataset": "iWildCam", "metrics": {"acc1": 0.21609684279404548, "acc5": 0.4210231123367063, "mean_per_class_recall": 0.16389162255648526, "acc_avg": 0.21791966259479523, "recall-macro_all": 0.16389162255648526, "F1-macro_all": 0.15011020079756165, "main_metric": 0.15011020079756165}}
{"key": "wilds/camelyon17", "dataset": "Camelyon17", "metrics": {"acc1": 0.7279022738495544, "acc5": null, "mean_per_class_recall": 0.7279022738495544, "acc_avg": 0.7279022932052612, "acc_slide:0": NaN, "count_slide:0": 0.0, "acc_slide:1": NaN, "count_slide:1": 0.0, "acc_slide:2": NaN, "count_slide:2": 0.0, "acc_slide:3": NaN, "count_slide:3": 0.0, "acc_slide:4": NaN, "count_slide:4": 0.0, "acc_slide:5": NaN, "count_slide:5": 0.0, "acc_slide:6": NaN, "count_slide:6": 0.0, "acc_slide:7": NaN, "count_slide:7": 0.0, "acc_slide:8": NaN, "count_slide:8": 0.0, "acc_slide:9": NaN, "count_slide:9": 0.0, "acc_slide:10": NaN, "count_slide:10": 0.0, "acc_slide:11": NaN, "count_slide:11": 0.0, "acc_slide:12": NaN, "count_slide:12": 0.0, "acc_slide:13": NaN, "count_slide:13": 0.0, "acc_slide:14": NaN, "count_slide:14": 0.0, "acc_slide:15": NaN, "count_slide:15": 0.0, "acc_slide:16": NaN, "count_slide:16": 0.0, "acc_slide:17": NaN, "count_slide:17": 0.0, "acc_slide:18": NaN, "count_slide:18": 0.0, "acc_slide:19": NaN, "count_slide:19": 0.0, "acc_slide:20": 0.9283464550971985, "count_slide:20": 3810.0, "acc_slide:21": 0.8716838359832764, "count_slide:21": 3694.0, "acc_slide:22": 0.6826629638671875, "count_slide:22": 7210.0, "acc_slide:23": 0.5633509755134583, "count_slide:23": 5288.0, "acc_slide:24": 0.8445709943771362, "count_slide:24": 7727.0, "acc_slide:25": 0.8880941271781921, "count_slide:25": 4334.0, "acc_slide:26": 0.7693315744400024, "count_slide:26": 3815.0, "acc_slide:27": 0.75, "count_slide:27": 4556.0, "acc_slide:28": 0.6695526838302612, "count_slide:28": 31878.0, "acc_slide:29": 0.7206090092658997, "count_slide:29": 12742.0, "acc_wg": 0.5633509755134583, "main_metric": 0.7279022738495544}}
{"key": "wilds/fmow", "dataset": "FMoW", "metrics": {"acc1": 0.21539714130631446, "acc5": 0.5117604487063506, "mean_per_class_recall": 0.2129885499249436, "acc_avg": 0.21539713442325592, "acc_year:0": NaN, "count_year:0": 0.0, "acc_year:1": NaN, "count_year:1": 0.0, "acc_year:2": NaN, "count_year:2": 0.0, "acc_year:3": NaN, "count_year:3": 0.0, "acc_year:4": NaN, "count_year:4": 0.0, "acc_year:5": NaN, "count_year:5": 0.0, "acc_year:6": NaN, "count_year:6": 0.0, "acc_year:7": NaN, "count_year:7": 0.0, "acc_year:8": NaN, "count_year:8": 0.0, "acc_year:9": NaN, "count_year:9": 0.0, "acc_year:10": NaN, "count_year:10": 0.0, "acc_year:11": NaN, "count_year:11": 0.0, "acc_year:12": NaN, "count_year:12": 0.0, "acc_year:13": NaN, "count_year:13": 0.0, "acc_year:14": 0.2280218005180359, "count_year:14": 15959.0, "acc_year:15": 0.18263132870197296, "count_year:15": 6149.0, "acc_worst_year": 0.18263132870197296, "acc_region:0": 0.19040901958942413, "count_region:0": 4963.0, "acc_region:1": 0.23147831857204437, "count_region:1": 5858.0, "acc_region:2": 0.1392209827899933, "count_region:2": 2593.0, "acc_region:3": 0.23741276562213898, "count_region:3": 8024.0, "acc_region:4": 0.28828829526901245, "count_region:4": 666.0, "acc_region:5": 0.75, "count_region:5": 4.0, "acc_worst_region": 0.1392209827899933, "main_metric": 0.1392209827899933}}
{"key": "fairness/dollar_street", "dataset": "Dollar Street", "metrics": {"acc1": 0.525834998572652, "acc5": 0.7904653154439052, "mean_per_class_recall": 0.5608762529814217, "acc_top5_avg": 0.7904652953147888, "acc_top5_income_ds:0": 0.6203271150588989, "count_income_ds:0": 856.0, "acc_top5_income_ds:1": 0.7952488660812378, "count_income_ds:1": 884.0, "acc_top5_income_ds:2": 0.8435072302818298, "count_income_ds:2": 901.0, "acc_top5_income_ds:3": 0.8990719318389893, "count_income_ds:3": 862.0, "acc_top5_wg": 0.6203271150588989, "main_metric": 0.6203271150588989}}
{"key": "fairness/geode", "dataset": "GeoDE", "metrics": {"acc1": 0.9112748238308777, "acc5": 0.9938340807174888, "mean_per_class_recall": 0.9114425088759956, "acc_avg": 0.9112748503684998, "acc_region:0": 0.896450936794281, "count_region:0": 2395.0, "acc_region:1": 0.9134328365325928, "count_region:1": 2010.0, "acc_region:2": 0.9153339862823486, "count_region:2": 2126.0, "acc_region:3": 0.9070364832878113, "count_region:3": 1947.0, "acc_region:4": 0.9095048308372498, "count_region:4": 1757.0, "acc_region:5": 0.9263204336166382, "count_region:5": 2253.0, "acc_wg": 0.896450936794281, "main_metric": 0.896450936794281}}
{"key": "fairness/fairface", "dataset": "FairFace", "metrics": {"acc_race_avg": 0.8975716829299927, "acc_race_race_binary:0": 0.5534772276878357, "count_race_binary:0": 2085.0, "acc_race_race_binary:1": 0.9784643054008484, "count_race_binary:1": 8869.0, "acc_race_wg": 0.5534772276878357, "acc_gender_avg": 0.9280627965927124, "acc_gender_race_binary:0": 0.9462829828262329, "acc_gender_race_binary:1": 0.9237794280052185, "acc_gender_wg": 0.9237794280052185, "acc_age_avg": 0.4575497508049011, "acc_age_race_binary:0": 0.46954435110092163, "acc_age_race_binary:1": 0.45472994446754456, "acc_age_wg": 0.45472994446754456, "acc_gender_x_avg": 0.9280627965927124, "acc_gender_x_race:0_gender:0": 0.7897371649742126, "count_race:0_gender:0": 799.0, "acc_gender_x_race:0_gender:1": 0.9537648558616638, "count_race:0_gender:1": 757.0, "acc_gender_x_race:1_gender:0": 0.9242424368858337, "count_race:1_gender:0": 1122.0, "acc_gender_x_race:1_gender:1": 0.9719626307487488, "count_race:1_gender:1": 963.0, "acc_gender_x_race:2_gender:0": 0.9030544757843018, "count_race:2_gender:0": 753.0, "acc_gender_x_race:2_gender:1": 0.9659239649772644, "count_race:2_gender:1": 763.0, "acc_gender_x_race:3_gender:0": 0.9041613936424255, "count_race:3_gender:0": 793.0, "acc_gender_x_race:3_gender:1": 0.9795180559158325, "count_race:3_gender:1": 830.0, "acc_gender_x_race:4_gender:0": 0.9606395959854126, "count_race:4_gender:0": 813.0, "acc_gender_x_race:4_gender:1": 0.9747474789619446, "count_race:4_gender:1": 396.0, "acc_gender_x_race:5_gender:0": 0.8476190567016602, "count_race:5_gender:0": 735.0, "acc_gender_x_race:5_gender:1": 0.9867647290229797, "count_race:5_gender:1": 680.0, "acc_gender_x_race:6_gender:0": 0.8661518692970276, "count_race:6_gender:0": 777.0, "acc_gender_x_race:6_gender:1": 0.9818887710571289, "count_race:6_gender:1": 773.0, "acc_gender_x_wg": 0.7897371649742126, "toxicity_crime_avg": 0.06554683297872543, "toxicity_crime_race:0": 0.027634961530566216, "count_race:0": 1556.0, "toxicity_crime_race:1": 0.17410072684288025, "count_race:1": 2085.0, "toxicity_crime_race:2": 0.033641159534454346, "count_race:2": 1516.0, "toxicity_crime_race:3": 0.04313000664114952, "count_race:3": 1623.0, "toxicity_crime_race:4": 0.090157151222229, "count_race:4": 1209.0, "toxicity_crime_race:5": 0.024028267711400986, "count_race:5": 1415.0, "toxicity_crime_race:6": 0.030967742204666138, "count_race:6": 1550.0, "toxicity_crime_wg": 0.024028267711400986, "toxicity_nonhuman_avg": 0.0, "toxicity_nonhuman_race:0": 0.0, "toxicity_nonhuman_race:1": 0.0, "toxicity_nonhuman_race:2": 0.0, "toxicity_nonhuman_race:3": 0.0, "toxicity_nonhuman_race:4": 0.0, "toxicity_nonhuman_race:5": 0.0, "toxicity_nonhuman_race:6": 0.0, "toxicity_nonhuman_wg": 0.0, "main_metric": null}}
{"key": "fairness/utkface", "dataset": "UTKFace", "metrics": {"acc_race_avg": 0.8818293213844299, "acc_race_race_binary:0": 0.7640929222106934, "count_race_binary:0": 10076.0, "acc_race_race_binary:1": 0.9688853025436401, "count_race_binary:1": 13627.0, "acc_race_wg": 0.7640929222106934, "acc_gender_avg": 0.9459562301635742, "acc_gender_race_binary:0": 0.9598054885864258, "acc_gender_race_binary:1": 0.9357158541679382, "acc_gender_wg": 0.9357158541679382, "acc_age_avg": 0.48930513858795166, "acc_age_race_binary:0": 0.4749900698661804, "acc_age_race_binary:1": 0.49988991022109985, "acc_age_wg": 0.4749900698661804, "acc_gender_x_avg": 0.9459562301635742, "acc_gender_x_race:0_gender:0": 0.9603106379508972, "count_race:0_gender:0": 2318.0, "acc_gender_x_race:0_gender:1": 0.9710144996643066, "count_race:0_gender:1": 2208.0, "acc_gender_x_race:1_gender:0": 0.938823938369751, "count_race:1_gender:0": 5476.0, "acc_gender_x_race:1_gender:1": 0.9847826361656189, "count_race:1_gender:1": 4600.0, "acc_gender_x_race:2_gender:0": 0.9314462542533875, "count_race:2_gender:0": 2261.0, "acc_gender_x_race:2_gender:1": 0.9702450633049011, "count_race:2_gender:1": 1714.0, "acc_gender_x_race:3_gender:0": 0.8012698292732239, "count_race:3_gender:0": 1575.0, "acc_gender_x_race:3_gender:1": 0.96880042552948, "count_race:3_gender:1": 1859.0, "acc_gender_x_race:4_gender:0": 0.8513157963752747, "count_race:4_gender:0": 760.0, "acc_gender_x_race:4_gender:1": 0.9678111672401428, "count_race:4_gender:1": 932.0, "acc_gender_x_wg": 0.8012698292732239, "toxicity_crime_avg": 0.06277686357498169, "toxicity_crime_race:0": 0.056783031672239304, "count_race:0": 4526.0, "toxicity_crime_race:1": 0.08703850954771042, "count_race:1": 10076.0, "toxicity_crime_race:2": 0.04176100715994835, "count_race:2": 3975.0, "toxicity_crime_race:3": 0.01572510227560997, "count_race:3": 3434.0, "toxicity_crime_race:4": 0.07919621467590332, "count_race:4": 1692.0, "toxicity_crime_wg": 0.01572510227560997, "toxicity_nonhuman_avg": 0.0005484537687152624, "toxicity_nonhuman_race:0": 0.0, "toxicity_nonhuman_race:1": 0.0010917030740529299, "toxicity_nonhuman_race:2": 0.0, "toxicity_nonhuman_race:3": 0.00029120559338480234, "toxicity_nonhuman_race:4": 0.0005910165491513908, "toxicity_nonhuman_wg": 0.0, "main_metric": null}}