aipib commited on
Commit
672a0a3
·
verified ·
1 Parent(s): b75410b

Upload LightOnOcrForConditionalGeneration

Browse files
Files changed (2) hide show
  1. config.json +6 -5
  2. model.safetensors +1 -1
config.json CHANGED
@@ -5,7 +5,7 @@
5
  "dtype": "bfloat16",
6
  "eos_token_id": 151645,
7
  "image_token_id": 151655,
8
- "model_type": "mistral3",
9
  "multimodal_projector_bias": false,
10
  "pad_token_id": 151643,
11
  "projector_hidden_act": "gelu",
@@ -16,7 +16,9 @@
16
  ],
17
  "attention_bias": false,
18
  "attention_dropout": 0,
 
19
  "dtype": "bfloat16",
 
20
  "head_dim": 128,
21
  "hidden_act": "silu",
22
  "hidden_size": 1024,
@@ -58,12 +60,12 @@
58
  "num_attention_heads": 16,
59
  "num_hidden_layers": 28,
60
  "num_key_value_heads": 8,
 
61
  "rms_norm_eps": 1e-06,
62
  "rope_parameters": {
63
  "rope_theta": 1000000,
64
  "rope_type": "default"
65
  },
66
- "rope_theta": 1000000,
67
  "sliding_window": null,
68
  "tie_word_embeddings": true,
69
  "use_cache": true,
@@ -71,7 +73,7 @@
71
  "use_sliding_window": false,
72
  "vocab_size": 151936
73
  },
74
- "use_cache": false,
75
  "transformers_version": "5.0.1.dev0",
76
  "vision_config": {
77
  "attention_dropout": 0,
@@ -90,8 +92,7 @@
90
  "rope_parameters": {
91
  "rope_theta": 10000,
92
  "rope_type": "default"
93
- },
94
- "rope_theta": 10000
95
  },
96
  "vision_feature_layer": -1
97
  }
 
5
  "dtype": "bfloat16",
6
  "eos_token_id": 151645,
7
  "image_token_id": 151655,
8
+ "model_type": "lighton_ocr",
9
  "multimodal_projector_bias": false,
10
  "pad_token_id": 151643,
11
  "projector_hidden_act": "gelu",
 
16
  ],
17
  "attention_bias": false,
18
  "attention_dropout": 0,
19
+ "bos_token_id": null,
20
  "dtype": "bfloat16",
21
+ "eos_token_id": null,
22
  "head_dim": 128,
23
  "hidden_act": "silu",
24
  "hidden_size": 1024,
 
60
  "num_attention_heads": 16,
61
  "num_hidden_layers": 28,
62
  "num_key_value_heads": 8,
63
+ "pad_token_id": null,
64
  "rms_norm_eps": 1e-06,
65
  "rope_parameters": {
66
  "rope_theta": 1000000,
67
  "rope_type": "default"
68
  },
 
69
  "sliding_window": null,
70
  "tie_word_embeddings": true,
71
  "use_cache": true,
 
73
  "use_sliding_window": false,
74
  "vocab_size": 151936
75
  },
76
+ "tie_word_embeddings": true,
77
  "transformers_version": "5.0.1.dev0",
78
  "vision_config": {
79
  "attention_dropout": 0,
 
92
  "rope_parameters": {
93
  "rope_theta": 10000,
94
  "rope_type": "default"
95
+ }
 
96
  },
97
  "vision_feature_layer": -1
98
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1377d35589f3f369fd5dec47f46a3583d5940c0cf942731b72336bb3a208210c
3
  size 2011367496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d36d6f29e2edd3588d6ec54d592b9ff4f8c64f443de631a284e82ac032aa6d5
3
  size 2011367496