daidv1112 commited on
Commit
054d349
·
verified ·
1 Parent(s): 68dcbc3

Delete Time_MoE/50M_etth1

Browse files
Time_MoE/50M_etth1/config.json DELETED
@@ -1,38 +0,0 @@
1
- {
2
- "_name_or_path": "Maple728/TimeMoE-50M",
3
- "apply_aux_loss": true,
4
- "architectures": [
5
- "TimeMoeForPrediction"
6
- ],
7
- "attention_dropout": 0.0,
8
- "auto_map": {
9
- "AutoConfig": "Maple728/TimeMoE-50M--configuration_time_moe.TimeMoeConfig",
10
- "AutoModelForCausalLM": "Maple728/TimeMoE-50M--modeling_time_moe.TimeMoeForPrediction"
11
- },
12
- "hidden_act": "silu",
13
- "hidden_size": 384,
14
- "horizon_lengths": [
15
- 1,
16
- 8,
17
- 32,
18
- 64
19
- ],
20
- "initializer_range": 0.02,
21
- "input_size": 1,
22
- "intermediate_size": 1536,
23
- "max_position_embeddings": 4096,
24
- "model_type": "time_moe",
25
- "num_attention_heads": 12,
26
- "num_experts": 8,
27
- "num_experts_per_tok": 2,
28
- "num_hidden_layers": 12,
29
- "num_key_value_heads": 12,
30
- "rms_norm_eps": 1e-06,
31
- "rope_theta": 10000,
32
- "router_aux_loss_factor": 0.02,
33
- "tie_word_embeddings": false,
34
- "torch_dtype": "float32",
35
- "transformers_version": "4.40.1",
36
- "use_cache": true,
37
- "use_dense": false
38
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Time_MoE/50M_etth1/generation_config.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "transformers_version": "4.40.1"
4
- }
 
 
 
 
 
Time_MoE/50M_etth1/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:210bb16b1f3e6eb95a13659f5c086dbabf173227ba317d5e66bb0612cc14b04f
3
- size 453464416
 
 
 
 
Time_MoE/50M_etth1/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e02f922e065ca0b6117673c530a9239ca336c0929fceed6aa59a456a23c64bea
3
- size 5457