danielhanchen commited on
Commit
e711faf
·
verified ·
1 Parent(s): f99d2da

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +2 -2
config.json CHANGED
@@ -9,7 +9,7 @@
9
  },
10
  "bos_token_id": 163584,
11
  "torch_dtype": "bfloat16",
12
- "eos_token_id": 163585,
13
  "ignore_index": -100,
14
  "media_placeholder_token_id": 163605,
15
  "model_type": "kimi_k25",
@@ -72,7 +72,7 @@
72
  "torch_dtype": "bfloat16",
73
  "early_stopping": false,
74
  "encoder_no_repeat_ngram_size": 0,
75
- "eos_token_id": 163585,
76
  "ep_size": 1,
77
  "exponential_decay_length_penalty": null,
78
  "finetuning_task": null,
 
9
  },
10
  "bos_token_id": 163584,
11
  "torch_dtype": "bfloat16",
12
+ "eos_token_id": 163586,
13
  "ignore_index": -100,
14
  "media_placeholder_token_id": 163605,
15
  "model_type": "kimi_k25",
 
72
  "torch_dtype": "bfloat16",
73
  "early_stopping": false,
74
  "encoder_no_repeat_ngram_size": 0,
75
+ "eos_token_id": 163586,
76
  "ep_size": 1,
77
  "exponential_decay_length_penalty": null,
78
  "finetuning_task": null,