--- license: mit datasets: - PolyAI/banking77 language: - en base_model: - google-bert/bert-base-uncased metrics: - f1 --- # BERT finetuned on Banking 77 Open Intent Classification Dataset This is a BERT model finetuned on the [PolyAI/banking77](https://huggingface.co/datasets/PolyAI/banking77) dataset. ## Training Configuration * PRETRAINED_MODEL_NAME = "bert-base-uncased" * BATCH_SIZE = 128 * LR_PRETRAIN = 2e-5 * EPOCHS_PRETRAIN = 20 * DEVICE = torch.device("cuda:0" if torch.cuda.is_available() else "cpu") * MAX_LEN = 128, to truncate long sequences down to 128 tokens, or pad short sequences up to 128 tokens