pere commited on
Commit
060247f
·
1 Parent(s): 8feb8d0
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": ".",
3
  "architectures": [
4
  "XLMRobertaForMaskedLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "./",
3
  "architectures": [
4
  "XLMRobertaForMaskedLM"
5
  ],
events.out.tfevents.1664916736.t1v-n-b899420b-w-0.278543.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf1e85eff2353955c51a626f18500c01e7dec159facddefb6bb22529c33a40bd
3
+ size 40
events.out.tfevents.1664917260.t1v-n-b899420b-w-0.281577.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ea148d2a1e3ef547244be3eea7fc176b551fc01c2e9700388697ed480fc3805
3
+ size 367912
events.out.tfevents.1664925833.t1v-n-b899420b-w-0.291698.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d3ad12a8242e1c39bbf79009716cdcff41d66500383a388c68874029bd8aaf7
3
+ size 1398158
events.out.tfevents.1664951079.t1v-n-b899420b-w-0.315648.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10a7aac4dddee5fdbd66befd292d0052ea2cc31baa879111903e5843f367d38c
3
+ size 40
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e176e8c732d80403ad8c03cdc305b5a1897e92ea407df51ef1cb427415e52299
3
+ size 1113187999
run.sh CHANGED
@@ -13,7 +13,7 @@ python run_mlm_flax_stream.py \
13
  --learning_rate="3e-4" \
14
  --warmup_steps="1000" \
15
  --overwrite_output_dir \
16
- --num_train_steps="10000" \
17
  --adam_beta1="0.9" \
18
  --adam_beta2="0.98" \
19
  --logging_steps="500" \
 
13
  --learning_rate="3e-4" \
14
  --warmup_steps="1000" \
15
  --overwrite_output_dir \
16
+ --num_train_steps="250000" \
17
  --adam_beta1="0.9" \
18
  --adam_beta2="0.98" \
19
  --logging_steps="500" \
run_mlm_flax_stream.py CHANGED
@@ -628,13 +628,20 @@ if __name__ == "__main__":
628
  print("*** Printing debug info")
629
  print(training_args.output_dir)
630
  print(training_args.push_to_hub)
631
- breakpoint()
632
- model.save_pretrained(
633
- training_args.output_dir,
634
- params=params,
635
- push_to_hub=training_args.push_to_hub,
636
- commit_message=f"Saving weights and logs of step {step+1}",
637
- )
 
 
 
 
 
 
 
638
 
639
  # update tqdm bar
640
  steps.update(1)
 
628
  print("*** Printing debug info")
629
  print(training_args.output_dir)
630
  print(training_args.push_to_hub)
631
+ try:
632
+ model.save_pretrained(
633
+ training_args.output_dir,
634
+ params=params,
635
+ push_to_hub=training_args.push_to_hub,
636
+ commit_message=f"Saving weights and logs of step {step+1}",
637
+ )
638
+ except:
639
+
640
+ model.save_pretrained(
641
+ training_args.output_dir,
642
+ params=params
643
+ )
644
+ print("probems pushing")
645
 
646
  # update tqdm bar
647
  steps.update(1)