{ "architectures": [ "ViTForImageClassification" ], "attention_probs_dropout_prob": 0.0, "dtype": "float32", "encoder_stride": 16, "hidden_act": "gelu", "hidden_dropout_prob": 0.0, "hidden_size": 768, "id2label": { "0": "0", "1": "01_ka", "10": "1", "11": "10_na(02)", "12": "11_ta(01)", "13": "12_tha(01)", "14": "13_da(01)", "15": "14_dha(01)", "16": "15_na(03)", "17": "16_ta(02)", "18": "17_tha(02)", "19": "18_da(02)", "2": "02_kha", "20": "19_dha(02)", "21": "2", "22": "20_na(04)", "23": "21_pa", "24": "22_pha", "25": "23_ba", "26": "24_bha", "27": "25_ma", "28": "26_ya", "29": "27_ra", "3": "03_ga", "30": "28_la(01)", "31": "29_va", "32": "3", "33": "30_la(02)", "34": "31_sa(01)", "35": "32_sa(02)", "36": "33_sa(03)", "37": "34_ha", "38": "4", "39": "5", "4": "04_gha", "40": "6", "41": "7", "42": "8", "43": "9", "5": "05_na(01)", "6": "06_ca", "7": "07_cha", "8": "08_ja", "9": "09_jha" }, "image_size": 224, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "0": "0", "01_ka": "1", "02_kha": "2", "03_ga": "3", "04_gha": "4", "05_na(01)": "5", "06_ca": "6", "07_cha": "7", "08_ja": "8", "09_jha": "9", "1": "10", "10_na(02)": "11", "11_ta(01)": "12", "12_tha(01)": "13", "13_da(01)": "14", "14_dha(01)": "15", "15_na(03)": "16", "16_ta(02)": "17", "17_tha(02)": "18", "18_da(02)": "19", "19_dha(02)": "20", "2": "21", "20_na(04)": "22", "21_pa": "23", "22_pha": "24", "23_ba": "25", "24_bha": "26", "25_ma": "27", "26_ya": "28", "27_ra": "29", "28_la(01)": "30", "29_va": "31", "3": "32", "30_la(02)": "33", "31_sa(01)": "34", "32_sa(02)": "35", "33_sa(03)": "36", "34_ha": "37", "4": "38", "5": "39", "6": "40", "7": "41", "8": "42", "9": "43" }, "layer_norm_eps": 1e-12, "model_type": "vit", "num_attention_heads": 12, "num_channels": 3, "num_hidden_layers": 12, "patch_size": 16, "pooler_act": "tanh", "pooler_output_size": 768, "problem_type": "single_label_classification", "qkv_bias": true, "transformers_version": "4.57.6" }