xfcxcxcdfdfd commited on
Commit
5f31828
·
verified ·
1 Parent(s): 779e16b

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,8 +1,7 @@
1
  {
2
- "_attn_implementation_autoset": true,
3
  "_name_or_path": "lilmeaty/testing_semifinal",
4
  "architectures": [
5
- "LlamaForCausalLM"
6
  ],
7
  "attention_bias": false,
8
  "attention_dropout": 0.0,
@@ -35,7 +34,7 @@
35
  },
36
  "rope_theta": 500000.0,
37
  "tie_word_embeddings": true,
38
- "torch_dtype": "float16",
39
  "transformers_version": "4.47.0",
40
  "use_cache": true,
41
  "vocab_size": 128256
 
1
  {
 
2
  "_name_or_path": "lilmeaty/testing_semifinal",
3
  "architectures": [
4
+ "PartiallySharedLayerModel"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
 
34
  },
35
  "rope_theta": 500000.0,
36
  "tie_word_embeddings": true,
37
+ "torch_dtype": "float32",
38
  "transformers_version": "4.47.0",
39
  "use_cache": true,
40
  "vocab_size": 128256
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 128000,
4
+ "eos_token_id": [
5
+ 128001,
6
+ 128008,
7
+ 128009
8
+ ],
9
+ "transformers_version": "4.47.0"
10
+ }
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ababbc023c6adae7e8ed6a34fa4effb2dc6258333a823a34ab6c10e7a46508f
3
+ size 4838750476
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd846cda54994ae32f75c0becad4dd838f25eadd1dcac8553e20ffddeef17070
3
+ size 4964730768
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:272bdf83f181c1e1077065b05c1dec1194530b08cb5a9bf5066437c0b50ff74d
3
+ size 679702024
model.safetensors.index.json ADDED
@@ -0,0 +1,206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 10483163276
4
+ },
5
+ "weight_map": {
6
+ "ability_weights": "model-00001-of-00003.safetensors",
7
+ "ability_weights_v2": "model-00001-of-00003.safetensors",
8
+ "accountability_head.bias": "model-00002-of-00003.safetensors",
9
+ "accountability_head.weight": "model-00002-of-00003.safetensors",
10
+ "action_head.bias": "model-00002-of-00003.safetensors",
11
+ "action_head.weight": "model-00002-of-00003.safetensors",
12
+ "adaptability_head.bias": "model-00002-of-00003.safetensors",
13
+ "adaptability_head.weight": "model-00002-of-00003.safetensors",
14
+ "agreeableness_head.bias": "model-00002-of-00003.safetensors",
15
+ "agreeableness_head.weight": "model-00002-of-00003.safetensors",
16
+ "altruism_head.bias": "model-00002-of-00003.safetensors",
17
+ "altruism_head.weight": "model-00002-of-00003.safetensors",
18
+ "assessment_head.bias": "model-00002-of-00003.safetensors",
19
+ "assessment_head.weight": "model-00002-of-00003.safetensors",
20
+ "capacity_head.bias": "model-00002-of-00003.safetensors",
21
+ "capacity_head.weight": "model-00002-of-00003.safetensors",
22
+ "capacity_predictions": "model-00001-of-00003.safetensors",
23
+ "civic_engagement_head.bias": "model-00003-of-00003.safetensors",
24
+ "civic_engagement_head.weight": "model-00003-of-00003.safetensors",
25
+ "cognitive_head.bias": "model-00002-of-00003.safetensors",
26
+ "cognitive_head.weight": "model-00002-of-00003.safetensors",
27
+ "collaboration_head.bias": "model-00002-of-00003.safetensors",
28
+ "collaboration_head.weight": "model-00002-of-00003.safetensors",
29
+ "communication_head.bias": "model-00002-of-00003.safetensors",
30
+ "communication_head.weight": "model-00002-of-00003.safetensors",
31
+ "compassion_head.bias": "model-00003-of-00003.safetensors",
32
+ "compassion_head.weight": "model-00003-of-00003.safetensors",
33
+ "conscientiousness_head.bias": "model-00002-of-00003.safetensors",
34
+ "conscientiousness_head.weight": "model-00002-of-00003.safetensors",
35
+ "courage_head.bias": "model-00002-of-00003.safetensors",
36
+ "courage_head.weight": "model-00002-of-00003.safetensors",
37
+ "creativity_head.bias": "model-00002-of-00003.safetensors",
38
+ "creativity_head.weight": "model-00002-of-00003.safetensors",
39
+ "critical_thinking_head.bias": "model-00002-of-00003.safetensors",
40
+ "critical_thinking_head.weight": "model-00002-of-00003.safetensors",
41
+ "cultural_sensitivity_head.bias": "model-00003-of-00003.safetensors",
42
+ "cultural_sensitivity_head.weight": "model-00003-of-00003.safetensors",
43
+ "curiosity_head.bias": "model-00002-of-00003.safetensors",
44
+ "curiosity_head.weight": "model-00002-of-00003.safetensors",
45
+ "decision_head.bias": "model-00002-of-00003.safetensors",
46
+ "decision_head.weight": "model-00002-of-00003.safetensors",
47
+ "digital_literacy_head.bias": "model-00003-of-00003.safetensors",
48
+ "digital_literacy_head.weight": "model-00003-of-00003.safetensors",
49
+ "draft_model.embed_tokens.weight": "model-00002-of-00003.safetensors",
50
+ "draft_model.emotion_vector": "model-00001-of-00003.safetensors",
51
+ "draft_model.lm_head.weight": "model-00002-of-00003.safetensors",
52
+ "draft_model.memory_keys": "model-00001-of-00003.safetensors",
53
+ "draft_model.memory_values": "model-00001-of-00003.safetensors",
54
+ "draft_model.norm.bias": "model-00002-of-00003.safetensors",
55
+ "draft_model.norm.weight": "model-00002-of-00003.safetensors",
56
+ "draft_model.prediction_head.bias": "model-00002-of-00003.safetensors",
57
+ "draft_model.prediction_head.weight": "model-00002-of-00003.safetensors",
58
+ "draft_model.response_selector.bias": "model-00002-of-00003.safetensors",
59
+ "draft_model.response_selector.weight": "model-00002-of-00003.safetensors",
60
+ "draft_model.shared_layer.input_layernorm.weight": "model-00001-of-00003.safetensors",
61
+ "draft_model.shared_layer.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
62
+ "draft_model.shared_layer.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
63
+ "draft_model.shared_layer.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
64
+ "draft_model.shared_layer.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
65
+ "draft_model.shared_layer.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
66
+ "draft_model.shared_layer.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
67
+ "draft_model.shared_layer.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
68
+ "draft_model.shared_layer.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
69
+ "draft_model.simulation_head.bias": "model-00002-of-00003.safetensors",
70
+ "draft_model.simulation_head.weight": "model-00002-of-00003.safetensors",
71
+ "draft_model.stimuli_vector": "model-00001-of-00003.safetensors",
72
+ "draft_model.unique_layer.input_layernorm.weight": "model-00001-of-00003.safetensors",
73
+ "draft_model.unique_layer.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
74
+ "draft_model.unique_layer.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
75
+ "draft_model.unique_layer.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
76
+ "draft_model.unique_layer.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
77
+ "draft_model.unique_layer.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
78
+ "draft_model.unique_layer.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
79
+ "draft_model.unique_layer.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
80
+ "draft_model.unique_layer.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
81
+ "embed_tokens.weight": "model-00001-of-00003.safetensors",
82
+ "emotion_vector": "model-00001-of-00003.safetensors",
83
+ "emotional_head.bias": "model-00002-of-00003.safetensors",
84
+ "emotional_head.weight": "model-00002-of-00003.safetensors",
85
+ "emotional_regulation_head.bias": "model-00003-of-00003.safetensors",
86
+ "emotional_regulation_head.weight": "model-00003-of-00003.safetensors",
87
+ "empathy_head.bias": "model-00002-of-00003.safetensors",
88
+ "empathy_head.weight": "model-00002-of-00003.safetensors",
89
+ "environmental_awareness_head.bias": "model-00003-of-00003.safetensors",
90
+ "environmental_awareness_head.weight": "model-00003-of-00003.safetensors",
91
+ "extraversion_head.bias": "model-00002-of-00003.safetensors",
92
+ "extraversion_head.weight": "model-00002-of-00003.safetensors",
93
+ "fairness_head.bias": "model-00002-of-00003.safetensors",
94
+ "fairness_head.weight": "model-00002-of-00003.safetensors",
95
+ "feedback_head.bias": "model-00002-of-00003.safetensors",
96
+ "feedback_head.weight": "model-00002-of-00003.safetensors",
97
+ "financial_literacy_head.bias": "model-00003-of-00003.safetensors",
98
+ "financial_literacy_head.weight": "model-00003-of-00003.safetensors",
99
+ "foresight_head.bias": "model-00002-of-00003.safetensors",
100
+ "foresight_head.weight": "model-00002-of-00003.safetensors",
101
+ "forgiveness_head.bias": "model-00003-of-00003.safetensors",
102
+ "forgiveness_head.weight": "model-00003-of-00003.safetensors",
103
+ "goal_head.bias": "model-00002-of-00003.safetensors",
104
+ "goal_head.weight": "model-00002-of-00003.safetensors",
105
+ "gratitude_head.bias": "model-00003-of-00003.safetensors",
106
+ "gratitude_head.weight": "model-00003-of-00003.safetensors",
107
+ "health_consciousness_head.bias": "model-00003-of-00003.safetensors",
108
+ "health_consciousness_head.weight": "model-00003-of-00003.safetensors",
109
+ "honesty_head.bias": "model-00002-of-00003.safetensors",
110
+ "honesty_head.weight": "model-00002-of-00003.safetensors",
111
+ "humility_head.bias": "model-00003-of-00003.safetensors",
112
+ "humility_head.weight": "model-00003-of-00003.safetensors",
113
+ "imagination_head.bias": "model-00002-of-00003.safetensors",
114
+ "imagination_head.weight": "model-00002-of-00003.safetensors",
115
+ "intuition_head.bias": "model-00002-of-00003.safetensors",
116
+ "intuition_head.weight": "model-00002-of-00003.safetensors",
117
+ "kindness_head.bias": "model-00002-of-00003.safetensors",
118
+ "kindness_head.weight": "model-00002-of-00003.safetensors",
119
+ "leadership_head.bias": "model-00002-of-00003.safetensors",
120
+ "leadership_head.weight": "model-00002-of-00003.safetensors",
121
+ "learning_head.bias": "model-00002-of-00003.safetensors",
122
+ "learning_head.weight": "model-00002-of-00003.safetensors",
123
+ "lifelong_learning_head.bias": "model-00003-of-00003.safetensors",
124
+ "lifelong_learning_head.weight": "model-00003-of-00003.safetensors",
125
+ "lm_head.weight": "model-00001-of-00003.safetensors",
126
+ "memory_keys": "model-00001-of-00003.safetensors",
127
+ "memory_values": "model-00001-of-00003.safetensors",
128
+ "moral_reasoning_head.bias": "model-00002-of-00003.safetensors",
129
+ "moral_reasoning_head.weight": "model-00002-of-00003.safetensors",
130
+ "morality_head.bias": "model-00002-of-00003.safetensors",
131
+ "morality_head.weight": "model-00002-of-00003.safetensors",
132
+ "negotiation_head.bias": "model-00002-of-00003.safetensors",
133
+ "negotiation_head.weight": "model-00002-of-00003.safetensors",
134
+ "neuroticism_head.bias": "model-00002-of-00003.safetensors",
135
+ "neuroticism_head.weight": "model-00002-of-00003.safetensors",
136
+ "norm.bias": "model-00001-of-00003.safetensors",
137
+ "norm.weight": "model-00001-of-00003.safetensors",
138
+ "novelty_vector": "model-00001-of-00003.safetensors",
139
+ "openness_head.bias": "model-00002-of-00003.safetensors",
140
+ "openness_head.weight": "model-00002-of-00003.safetensors",
141
+ "optimism_head.bias": "model-00003-of-00003.safetensors",
142
+ "optimism_head.weight": "model-00003-of-00003.safetensors",
143
+ "patience_head.bias": "model-00002-of-00003.safetensors",
144
+ "patience_head.weight": "model-00002-of-00003.safetensors",
145
+ "perseverance_head.bias": "model-00002-of-00003.safetensors",
146
+ "perseverance_head.weight": "model-00002-of-00003.safetensors",
147
+ "planning_head.bias": "model-00002-of-00003.safetensors",
148
+ "planning_head.weight": "model-00002-of-00003.safetensors",
149
+ "prediction_head.bias": "model-00001-of-00003.safetensors",
150
+ "prediction_head.weight": "model-00001-of-00003.safetensors",
151
+ "problem_solving_head.bias": "model-00002-of-00003.safetensors",
152
+ "problem_solving_head.weight": "model-00002-of-00003.safetensors",
153
+ "questioning_head.bias": "model-00002-of-00003.safetensors",
154
+ "questioning_head.weight": "model-00002-of-00003.safetensors",
155
+ "reflection_head.bias": "model-00002-of-00003.safetensors",
156
+ "reflection_head.weight": "model-00002-of-00003.safetensors",
157
+ "resilience_head.bias": "model-00002-of-00003.safetensors",
158
+ "resilience_head.weight": "model-00002-of-00003.safetensors",
159
+ "response_selector.bias": "model-00001-of-00003.safetensors",
160
+ "response_selector.weight": "model-00001-of-00003.safetensors",
161
+ "self_awareness_head.bias": "model-00003-of-00003.safetensors",
162
+ "self_awareness_head.weight": "model-00003-of-00003.safetensors",
163
+ "self_awareness_vector": "model-00001-of-00003.safetensors",
164
+ "self_control_head.bias": "model-00002-of-00003.safetensors",
165
+ "self_control_head.weight": "model-00002-of-00003.safetensors",
166
+ "self_efficacy_head.bias": "model-00003-of-00003.safetensors",
167
+ "self_efficacy_head.weight": "model-00003-of-00003.safetensors",
168
+ "self_esteem_head.bias": "model-00003-of-00003.safetensors",
169
+ "self_esteem_head.weight": "model-00003-of-00003.safetensors",
170
+ "self_improvement_head.bias": "model-00002-of-00003.safetensors",
171
+ "self_improvement_head.weight": "model-00002-of-00003.safetensors",
172
+ "shaman_head.bias": "model-00002-of-00003.safetensors",
173
+ "shaman_head.weight": "model-00002-of-00003.safetensors",
174
+ "shared_layer.input_layernorm.weight": "model-00001-of-00003.safetensors",
175
+ "shared_layer.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
176
+ "shared_layer.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
177
+ "shared_layer.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
178
+ "shared_layer.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
179
+ "shared_layer.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
180
+ "shared_layer.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
181
+ "shared_layer.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
182
+ "shared_layer.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
183
+ "simulation_head.bias": "model-00001-of-00003.safetensors",
184
+ "simulation_head.weight": "model-00001-of-00003.safetensors",
185
+ "stimuli_vector": "model-00001-of-00003.safetensors",
186
+ "strategy_head.bias": "model-00002-of-00003.safetensors",
187
+ "strategy_head.weight": "model-00002-of-00003.safetensors",
188
+ "stress_management_head.bias": "model-00003-of-00003.safetensors",
189
+ "stress_management_head.weight": "model-00003-of-00003.safetensors",
190
+ "sustainability_head.bias": "model-00002-of-00003.safetensors",
191
+ "sustainability_head.weight": "model-00002-of-00003.safetensors",
192
+ "time_management_head.bias": "model-00003-of-00003.safetensors",
193
+ "time_management_head.weight": "model-00003-of-00003.safetensors",
194
+ "unique_layer.input_layernorm.weight": "model-00001-of-00003.safetensors",
195
+ "unique_layer.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
196
+ "unique_layer.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
197
+ "unique_layer.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
198
+ "unique_layer.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
199
+ "unique_layer.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
200
+ "unique_layer.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
201
+ "unique_layer.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
202
+ "unique_layer.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
203
+ "wisdom_head.bias": "model-00002-of-00003.safetensors",
204
+ "wisdom_head.weight": "model-00002-of-00003.safetensors"
205
+ }
206
+ }