pere commited on
Commit
fe5f6e7
·
1 Parent(s): 2349a63

copies from roberta xlm

Browse files
Files changed (4) hide show
  1. flax_model.msgpack +0 -3
  2. generate_pt_model.py +0 -6
  3. run.sh +0 -21
  4. tokenizer.json +2 -2
flax_model.msgpack DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c923408bb46e4dc6eadd451c0875742de42256d53079fe7fe233c14e6278f48
3
- size 1113187999
 
 
 
 
generate_pt_model.py DELETED
@@ -1,6 +0,0 @@
1
- from transformers import XLMRobertaForMaskedLM, XLMRobertaConfig
2
- config = XLMRobertaConfig.from_pretrained("./")
3
- model = XLMRobertaForMaskedLM.from_pretrained("./",config=config,from_flax=True)
4
- model.save_pretrained("./")
5
-
6
-
 
 
 
 
 
 
 
run.sh DELETED
@@ -1,21 +0,0 @@
1
- python run_mlm_flax_stream.py \
2
- --output_dir="../roberta-base-scand-xlm" \
3
- --model_name_or_path="xlm-roberta-base" \
4
- --config_name="./" \
5
- --tokenizer_name="./" \
6
- --dataset_name="NbAiLab/scandinavian" \
7
- --max_seq_length="512" \
8
- --weight_decay="0.01" \
9
- --per_device_train_batch_size="62" \
10
- --per_device_eval_batch_size="62" \
11
- --learning_rate="3e-4" \
12
- --warmup_steps="50000" \
13
- --overwrite_output_dir \
14
- --num_train_steps="1000000" \
15
- --adam_beta1="0.9" \
16
- --adam_beta2="0.98" \
17
- --logging_steps="5000" \
18
- --save_steps="25000" \
19
- --eval_steps="25000" \
20
- --dtype="bfloat16" \
21
- --push_to_hub
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:62c24cdc13d4c9952d63718d6c9fa4c287974249e16b7ade6d5a85e7bbb75626
3
- size 17082660
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a898ea75433890f6610f4e470b8ebeb0c21dce5c8dd61f892eb09eb5919d2e2c
3
+ size 9096718