Training in progress, epoch 1
Browse files- .gitattributes +1 -0
- .gitignore +1 -0
- config.json +31 -0
- pytorch_model.bin +3 -0
- runs/Sep04_17-03-00_5cce0c882db2/1662311183.756341/events.out.tfevents.1662311183.5cce0c882db2.17.1 +3 -0
- runs/Sep04_17-03-00_5cce0c882db2/events.out.tfevents.1662311183.5cce0c882db2.17.0 +3 -0
- runs/Sep04_17-14-28_5cce0c882db2/1662311676.7825725/events.out.tfevents.1662311676.5cce0c882db2.17.3 +3 -0
- runs/Sep04_17-14-28_5cce0c882db2/events.out.tfevents.1662311676.5cce0c882db2.17.2 +3 -0
- sentencepiece.bpe.model +3 -0
- special_tokens_map.json +9 -0
- tokenizer.json +3 -0
- tokenizer_config.json +23 -0
- training_args.bin +3 -0
.gitattributes
CHANGED
|
@@ -30,3 +30,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 30 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 31 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 32 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 30 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 31 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 32 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
.gitignore
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
checkpoint-*/
|
config.json
ADDED
|
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "deepset/xlm-roberta-base-squad2",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"XLMRobertaForQuestionAnswering"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"bos_token_id": 0,
|
| 8 |
+
"classifier_dropout": null,
|
| 9 |
+
"eos_token_id": 2,
|
| 10 |
+
"gradient_checkpointing": false,
|
| 11 |
+
"hidden_act": "gelu",
|
| 12 |
+
"hidden_dropout_prob": 0.1,
|
| 13 |
+
"hidden_size": 768,
|
| 14 |
+
"initializer_range": 0.02,
|
| 15 |
+
"intermediate_size": 3072,
|
| 16 |
+
"language": "english",
|
| 17 |
+
"layer_norm_eps": 1e-05,
|
| 18 |
+
"max_position_embeddings": 514,
|
| 19 |
+
"model_type": "xlm-roberta",
|
| 20 |
+
"name": "XLMRoberta",
|
| 21 |
+
"num_attention_heads": 12,
|
| 22 |
+
"num_hidden_layers": 12,
|
| 23 |
+
"output_past": true,
|
| 24 |
+
"pad_token_id": 1,
|
| 25 |
+
"position_embedding_type": "absolute",
|
| 26 |
+
"torch_dtype": "float32",
|
| 27 |
+
"transformers_version": "4.20.1",
|
| 28 |
+
"type_vocab_size": 1,
|
| 29 |
+
"use_cache": true,
|
| 30 |
+
"vocab_size": 250002
|
| 31 |
+
}
|
pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fb0a3ca6d4b4f5d2db3b5ba62d806f7420b543b1d35aebf23fa27bef88cedd61
|
| 3 |
+
size 1109889457
|
runs/Sep04_17-03-00_5cce0c882db2/1662311183.756341/events.out.tfevents.1662311183.5cce0c882db2.17.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:137ec522d388676f69b4e7b6a296d75383dbd39a35c9728b3874ab469cfe1255
|
| 3 |
+
size 5357
|
runs/Sep04_17-03-00_5cce0c882db2/events.out.tfevents.1662311183.5cce0c882db2.17.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ba9ccc7a1d19871b1ed8cbbc1605c94a1d666f5c192ccfbd128cac7e68a53cf6
|
| 3 |
+
size 3946
|
runs/Sep04_17-14-28_5cce0c882db2/1662311676.7825725/events.out.tfevents.1662311676.5cce0c882db2.17.3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:87132070814e2cea6e951977feab61549b8ce1c235ae8667efd8f383a7fd895a
|
| 3 |
+
size 5357
|
runs/Sep04_17-14-28_5cce0c882db2/events.out.tfevents.1662311676.5cce0c882db2.17.2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9f7263a01c469d6122e0d6c04c63d32180b9a426ceab48a4f8f442c933d98e4b
|
| 3 |
+
size 3789
|
sentencepiece.bpe.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
|
| 3 |
+
size 5069051
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": "<s>",
|
| 3 |
+
"cls_token": "<s>",
|
| 4 |
+
"eos_token": "</s>",
|
| 5 |
+
"mask_token": "<mask>",
|
| 6 |
+
"pad_token": "<pad>",
|
| 7 |
+
"sep_token": "</s>",
|
| 8 |
+
"unk_token": "<unk>"
|
| 9 |
+
}
|
tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c733801366717da20ca45ad8140c639146baed31ad67db0586d9d3eace799532
|
| 3 |
+
size 17098346
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": "<s>",
|
| 3 |
+
"cls_token": "<s>",
|
| 4 |
+
"do_lower_case": false,
|
| 5 |
+
"eos_token": "</s>",
|
| 6 |
+
"full_tokenizer_file": null,
|
| 7 |
+
"mask_token": {
|
| 8 |
+
"__type": "AddedToken",
|
| 9 |
+
"content": "<mask>",
|
| 10 |
+
"lstrip": true,
|
| 11 |
+
"normalized": true,
|
| 12 |
+
"rstrip": false,
|
| 13 |
+
"single_word": false
|
| 14 |
+
},
|
| 15 |
+
"model_max_length": 512,
|
| 16 |
+
"name_or_path": "deepset/xlm-roberta-base-squad2",
|
| 17 |
+
"pad_token": "<pad>",
|
| 18 |
+
"sep_token": "</s>",
|
| 19 |
+
"sp_model_kwargs": {},
|
| 20 |
+
"special_tokens_map_file": "/root/.cache/huggingface/transformers/b285d730ad6193f7e96b41152fbc5fb05620f0961f6f513e2d36868dfb1eb2a7.0dc5b1041f62041ebbd23b1297f2f573769d5c97d8b7c28180ec86b8f6185aa8",
|
| 21 |
+
"tokenizer_class": "XLMRobertaTokenizer",
|
| 22 |
+
"unk_token": "<unk>"
|
| 23 |
+
}
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:127a1cc84a8e64272c7e92e3f682cd54ba6fbab7c214df387f45d8ec1eacf450
|
| 3 |
+
size 3311
|