robinfaro commited on
Commit
f2f59d0
·
verified ·
1 Parent(s): 88d6e74

Upload MoEGPTForCausalLM

Browse files
Files changed (3) hide show
  1. config.json +20 -79
  2. generation_config.json +4 -0
  3. model.safetensors +3 -0
config.json CHANGED
@@ -1,89 +1,30 @@
1
  {
2
- "return_dict": true,
3
- "output_hidden_states": false,
4
- "output_attentions": false,
5
- "torchscript": false,
6
- "torch_dtype": null,
7
- "use_bfloat16": false,
8
- "tf_legacy_loss": false,
9
- "pruned_heads": {},
10
- "tie_word_embeddings": true,
11
- "chunk_size_feed_forward": 0,
12
- "is_encoder_decoder": false,
13
- "is_decoder": false,
14
- "cross_attention_hidden_size": null,
15
- "add_cross_attention": false,
16
- "tie_encoder_decoder": false,
17
- "max_length": 20,
18
- "min_length": 0,
19
- "do_sample": false,
20
- "early_stopping": false,
21
- "num_beams": 1,
22
- "num_beam_groups": 1,
23
- "diversity_penalty": 0.0,
24
- "temperature": 1.0,
25
- "top_k": 50,
26
- "top_p": 1.0,
27
- "typical_p": 1.0,
28
- "repetition_penalty": 1.0,
29
- "length_penalty": 1.0,
30
- "no_repeat_ngram_size": 0,
31
- "encoder_no_repeat_ngram_size": 0,
32
- "bad_words_ids": null,
33
- "num_return_sequences": 1,
34
- "output_scores": false,
35
- "return_dict_in_generate": false,
36
- "forced_bos_token_id": null,
37
- "forced_eos_token_id": null,
38
- "remove_invalid_values": false,
39
- "exponential_decay_length_penalty": null,
40
- "suppress_tokens": null,
41
- "begin_suppress_tokens": null,
42
  "architectures": [
43
  "MoEGPTForCausalLM"
44
  ],
45
- "finetuning_task": null,
46
- "id2label": {
47
- "0": "LABEL_0",
48
- "1": "LABEL_1"
49
- },
50
- "label2id": {
51
- "LABEL_0": 0,
52
- "LABEL_1": 1
53
  },
54
- "tokenizer_class": null,
55
- "prefix": null,
56
- "bos_token_id": null,
57
- "pad_token_id": null,
58
- "eos_token_id": null,
59
- "sep_token_id": null,
60
- "decoder_start_token_id": null,
61
- "task_specific_params": null,
62
- "problem_type": null,
63
- "_name_or_path": "",
64
- "_attn_implementation_autoset": false,
65
- "transformers_version": "4.51.3",
66
- "shared_attention": true,
67
- "vocab_size": 50304,
68
- "n_embd": 1152,
69
- "n_layer": 24,
70
- "n_head": 16,
71
- "sequence_length": 1024,
72
  "moe": false,
73
- "moe_routing": null,
74
  "moe_num_experts": 1,
75
  "moe_num_experts_per_tok": 2,
76
- "moe_softmax_order": "softmax_topk",
77
  "moe_router_loss": "load_balancing_z_loss",
78
- "moe_aux_loss_factor": 0.01,
 
79
  "moe_z_loss_factor": 1.0,
80
- "mlp_dim_exp_factor": 1.0,
81
- "dropout": 0.0,
82
- "bias": false,
83
- "auto_map": {
84
- "AutoConfig": "configuration.MoEGPTConfig",
85
- "AutoModelForCausalLM": "modeling.MoEGPTForCausalLM",
86
- "AutoTokenizer": "GPT2TokenizerFast"
87
- },
88
- "model_type": "moegpt"
89
- }
 
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "architectures": [
3
  "MoEGPTForCausalLM"
4
  ],
5
+ "auto_map": {
6
+ "AutoConfig": "configuration.MoEGPTConfig",
7
+ "AutoModelForCausalLM": "modeling.MoEGPTForCausalLM",
8
+ "AutoTokenizer": "GPT2TokenizerFast"
 
 
 
 
9
  },
10
+ "bias": false,
11
+ "dropout": 0.0,
12
+ "mlp_dim_exp_factor": 1.0,
13
+ "model_type": "moegpt",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  "moe": false,
15
+ "moe_aux_loss_factor": 0.01,
16
  "moe_num_experts": 1,
17
  "moe_num_experts_per_tok": 2,
 
18
  "moe_router_loss": "load_balancing_z_loss",
19
+ "moe_routing": null,
20
+ "moe_softmax_order": "softmax_topk",
21
  "moe_z_loss_factor": 1.0,
22
+ "n_embd": 1152,
23
+ "n_head": 16,
24
+ "n_layer": 24,
25
+ "sequence_length": 1024,
26
+ "shared_attention": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.51.3",
29
+ "vocab_size": 50304
30
+ }
 
generation_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "transformers_version": "4.51.3"
4
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d7909678724591f642f5f44777ebd3817dcf7b4264ad0c2d7f8468bbf436297
3
+ size 1997385672