copies from roberta xlm
Browse files- flax_model.msgpack +0 -3
- generate_pt_model.py +0 -6
- run.sh +0 -21
- tokenizer.json +2 -2
flax_model.msgpack
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:4c923408bb46e4dc6eadd451c0875742de42256d53079fe7fe233c14e6278f48
|
| 3 |
-
size 1113187999
|
|
|
|
|
|
|
|
|
|
|
|
generate_pt_model.py
DELETED
|
@@ -1,6 +0,0 @@
|
|
| 1 |
-
from transformers import XLMRobertaForMaskedLM, XLMRobertaConfig
|
| 2 |
-
config = XLMRobertaConfig.from_pretrained("./")
|
| 3 |
-
model = XLMRobertaForMaskedLM.from_pretrained("./",config=config,from_flax=True)
|
| 4 |
-
model.save_pretrained("./")
|
| 5 |
-
|
| 6 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run.sh
DELETED
|
@@ -1,21 +0,0 @@
|
|
| 1 |
-
python run_mlm_flax_stream.py \
|
| 2 |
-
--output_dir="../roberta-base-scand-xlm" \
|
| 3 |
-
--model_name_or_path="xlm-roberta-base" \
|
| 4 |
-
--config_name="./" \
|
| 5 |
-
--tokenizer_name="./" \
|
| 6 |
-
--dataset_name="NbAiLab/scandinavian" \
|
| 7 |
-
--max_seq_length="512" \
|
| 8 |
-
--weight_decay="0.01" \
|
| 9 |
-
--per_device_train_batch_size="62" \
|
| 10 |
-
--per_device_eval_batch_size="62" \
|
| 11 |
-
--learning_rate="3e-4" \
|
| 12 |
-
--warmup_steps="50000" \
|
| 13 |
-
--overwrite_output_dir \
|
| 14 |
-
--num_train_steps="1000000" \
|
| 15 |
-
--adam_beta1="0.9" \
|
| 16 |
-
--adam_beta2="0.98" \
|
| 17 |
-
--logging_steps="5000" \
|
| 18 |
-
--save_steps="25000" \
|
| 19 |
-
--eval_steps="25000" \
|
| 20 |
-
--dtype="bfloat16" \
|
| 21 |
-
--push_to_hub
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
tokenizer.json
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a898ea75433890f6610f4e470b8ebeb0c21dce5c8dd61f892eb09eb5919d2e2c
|
| 3 |
+
size 9096718
|