pere commited on
Commit
e1806f2
·
1 Parent(s): 0a98a9d

Saving weights and logs of step 5000

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": ".",
3
  "architectures": [
4
  "RobertaForMaskedLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "./",
3
  "architectures": [
4
  "RobertaForMaskedLM"
5
  ],
events.out.tfevents.1638605041.t1v-n-1a0a7c50-w-0.197451.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9836328894d46fb754b13d1774fdfa87b83c7e4ffce927c5127599a81d52ce00
3
+ size 35676
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8a721cc30cc072ae6a0b03c30aa65698ef14b591e8618bac15b3fcda27750f1
3
+ size 498796983
run_flax.sh CHANGED
@@ -3,8 +3,7 @@
3
  --model_type="roberta" \
4
  --config_name="./" \
5
  --tokenizer_name="./" \
6
- --train_file="/mnt/disks/flaxdisk/smallcorpus/train-shard-0001-of-0001.json" \
7
- --validation_file="/mnt/disks/flaxdisk/smallcorpus/validation-shard-0001-of-0001.json" \
8
  --max_seq_length="128" \
9
  --weight_decay="0.01" \
10
  --per_device_train_batch_size="232" \
@@ -12,7 +11,7 @@
12
  --learning_rate="6e-4" \
13
  --warmup_steps="5000" \
14
  --overwrite_output_dir \
15
- --num_train_epochss="10" \
16
  --adam_beta1="0.9" \
17
  --adam_beta2="0.98" \
18
  --logging_steps="5000" \
 
3
  --model_type="roberta" \
4
  --config_name="./" \
5
  --tokenizer_name="./" \
6
+ --dataset_name="NbAiLab/NCC_small" \
 
7
  --max_seq_length="128" \
8
  --weight_decay="0.01" \
9
  --per_device_train_batch_size="232" \
 
11
  --learning_rate="6e-4" \
12
  --warmup_steps="5000" \
13
  --overwrite_output_dir \
14
+ --num_train_epochs="1000" \
15
  --adam_beta1="0.9" \
16
  --adam_beta2="0.98" \
17
  --logging_steps="5000" \