File size: 2,631 Bytes
6502387 8c84cb4 a2aabb8 6502387 3f7f53a a2aabb8 3f7f53a a2aabb8 3f7f53a a2aabb8 3f7f53a a2aabb8 3f7f53a 9530fb8 a2aabb8 9530fb8 a2aabb8 9530fb8 a2aabb8 9530fb8 a2aabb8 9530fb8 483c3e8 a2aabb8 483c3e8 a2aabb8 483c3e8 a2aabb8 483c3e8 a2aabb8 483c3e8 0c637d2 a2aabb8 0c637d2 a2aabb8 0c637d2 a2aabb8 0c637d2 a2aabb8 0c637d2 6502387 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 |
---
license: mit
tags:
- generated_from_trainer
datasets:
- squad_v2
- quoref
- adversarial_qa
- ibm/duorc
model-index:
- name: rob-base-superqa2
results:
- task:
type: question-answering
name: Question Answering
dataset:
name: squad_v2
type: squad_v2
config: squad_v2
split: validation
metrics:
- type: exact_match
value: 79.2365
name: Exact Match
verified: true
- type: f1
value: 82.3326
name: F1
verified: true
- task:
type: question-answering
name: Question Answering
dataset:
name: adversarial_qa
type: adversarial_qa
config: adversarialQA
split: test
metrics:
- type: exact_match
value: 12.4
name: Exact Match
verified: true
- type: f1
value: 12.4
name: F1
verified: true
- task:
type: question-answering
name: Question Answering
dataset:
name: adversarial_qa
type: adversarial_qa
config: adversarialQA
split: validation
metrics:
- type: exact_match
value: 42.3667
name: Exact Match
verified: true
- type: f1
value: 53.3255
name: F1
verified: true
- task:
type: question-answering
name: Question Answering
dataset:
name: squad
type: squad
config: plain_text
split: validation
metrics:
- type: exact_match
value: 86.1925
name: Exact Match
verified: true
- type: f1
value: 92.4306
name: F1
verified: true
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# rob-base-superqa2
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- distributed_type: multi-GPU
- num_devices: 8
- total_train_batch_size: 256
- total_eval_batch_size: 256
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 2.0
### Training results
### Framework versions
- Transformers 4.21.1
- Pytorch 1.11.0a0+gita4c10ee
- Datasets 2.4.0
- Tokenizers 0.12.1
|