{ "_name_or_path": "vinai/phobert-base", "architectures": [ "RobertaForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "am_thuc", "1": "bong_da", "2": "cho_thue", "3": "doi_song", "4": "dong_vat", "5": "mua_ban", "6": "nhac", "7": "phim", "8": "phu_kien", "9": "sach", "10": "showbiz", "11": "the_thao", "12": "thoi_trang_nam", "13": "thoi_trang_nu", "14": "thuc_vat", "15": "tin_bds", "16": "tin_tuc", "17": "tri_thuc" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "am_thuc": 0, "bong_da": 1, "cho_thue": 2, "doi_song": 3, "dong_vat": 4, "mua_ban": 5, "nhac": 6, "phim": 7, "phu_kien": 8, "sach": 9, "showbiz": 10, "the_thao": 11, "thoi_trang_nam": 12, "thoi_trang_nu": 13, "thuc_vat": 14, "tin_bds": 15, "tin_tuc": 16, "tri_thuc": 17 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 258, "model_type": "roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "position_embedding_type": "absolute", "problem_type": "multi_label_classification", "tokenizer_class": "PhobertTokenizer", "torch_dtype": "float32", "transformers_version": "4.19.2", "type_vocab_size": 1, "use_cache": true, "vocab_size": 64001 }