daidv1112 commited on
Commit
68dcbc3
·
verified ·
1 Parent(s): 7223a22

Delete Time_MoE/checkpoints/timerxl_finetuned

Browse files
Files changed (21) hide show
  1. Time_MoE/checkpoints/timerxl_finetuned/ETTh1/epoch-1/config.json +0 -27
  2. Time_MoE/checkpoints/timerxl_finetuned/ETTh1/epoch-1/generation_config.json +0 -4
  3. Time_MoE/checkpoints/timerxl_finetuned/ETTh1/epoch-1/model.safetensors +0 -3
  4. Time_MoE/checkpoints/timerxl_finetuned/ETTh2/epoch-1/config.json +0 -27
  5. Time_MoE/checkpoints/timerxl_finetuned/ETTh2/epoch-1/generation_config.json +0 -4
  6. Time_MoE/checkpoints/timerxl_finetuned/ETTh2/epoch-1/model.safetensors +0 -3
  7. Time_MoE/checkpoints/timerxl_finetuned/ETTm1/epoch-1/config.json +0 -27
  8. Time_MoE/checkpoints/timerxl_finetuned/ETTm1/epoch-1/generation_config.json +0 -4
  9. Time_MoE/checkpoints/timerxl_finetuned/ETTm1/epoch-1/model.safetensors +0 -3
  10. Time_MoE/checkpoints/timerxl_finetuned/ETTm2/epoch-1/config.json +0 -27
  11. Time_MoE/checkpoints/timerxl_finetuned/ETTm2/epoch-1/generation_config.json +0 -4
  12. Time_MoE/checkpoints/timerxl_finetuned/ETTm2/epoch-1/model.safetensors +0 -3
  13. Time_MoE/checkpoints/timerxl_finetuned/epoch-1/config.json +0 -27
  14. Time_MoE/checkpoints/timerxl_finetuned/epoch-1/generation_config.json +0 -4
  15. Time_MoE/checkpoints/timerxl_finetuned/epoch-1/model.safetensors +0 -3
  16. Time_MoE/checkpoints/timerxl_finetuned/epoch-2/config.json +0 -27
  17. Time_MoE/checkpoints/timerxl_finetuned/epoch-2/generation_config.json +0 -4
  18. Time_MoE/checkpoints/timerxl_finetuned/epoch-2/model.safetensors +0 -3
  19. Time_MoE/checkpoints/timerxl_finetuned/epoch-3/config.json +0 -27
  20. Time_MoE/checkpoints/timerxl_finetuned/epoch-3/generation_config.json +0 -4
  21. Time_MoE/checkpoints/timerxl_finetuned/epoch-3/model.safetensors +0 -3
Time_MoE/checkpoints/timerxl_finetuned/ETTh1/epoch-1/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "thuml/timer-base-84m",
3
- "architectures": [
4
- "TimerForPrediction"
5
- ],
6
- "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "thuml/timer-base-84m--configuration_timer.TimerConfig",
9
- "AutoModelForCausalLM": "thuml/timer-base-84m--modeling_timer.TimerForPrediction"
10
- },
11
- "hidden_act": "silu",
12
- "hidden_size": 1024,
13
- "initializer_range": 0.02,
14
- "input_token_len": 96,
15
- "intermediate_size": 2048,
16
- "max_position_embeddings": 10000,
17
- "model_type": "timer",
18
- "num_attention_heads": 8,
19
- "num_hidden_layers": 8,
20
- "output_token_lens": [
21
- 96
22
- ],
23
- "rope_theta": 10000,
24
- "torch_dtype": "float32",
25
- "transformers_version": "4.40.1",
26
- "use_cache": true
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTh1/epoch-1/generation_config.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "transformers_version": "4.40.1"
4
- }
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTh1/epoch-1/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa1e45adcb6501e41469450f606051d195938f3d6d0005230248732d2bf8066c
3
- size 336580760
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTh2/epoch-1/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "thuml/timer-base-84m",
3
- "architectures": [
4
- "TimerForPrediction"
5
- ],
6
- "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "thuml/timer-base-84m--configuration_timer.TimerConfig",
9
- "AutoModelForCausalLM": "thuml/timer-base-84m--modeling_timer.TimerForPrediction"
10
- },
11
- "hidden_act": "silu",
12
- "hidden_size": 1024,
13
- "initializer_range": 0.02,
14
- "input_token_len": 96,
15
- "intermediate_size": 2048,
16
- "max_position_embeddings": 10000,
17
- "model_type": "timer",
18
- "num_attention_heads": 8,
19
- "num_hidden_layers": 8,
20
- "output_token_lens": [
21
- 96
22
- ],
23
- "rope_theta": 10000,
24
- "torch_dtype": "float32",
25
- "transformers_version": "4.40.1",
26
- "use_cache": true
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTh2/epoch-1/generation_config.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "transformers_version": "4.40.1"
4
- }
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTh2/epoch-1/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c423352db7b072a809b08f66fb2b7400f5cd13c0382601c619b2d5e15c8bb8c8
3
- size 336580760
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTm1/epoch-1/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "thuml/timer-base-84m",
3
- "architectures": [
4
- "TimerForPrediction"
5
- ],
6
- "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "thuml/timer-base-84m--configuration_timer.TimerConfig",
9
- "AutoModelForCausalLM": "thuml/timer-base-84m--modeling_timer.TimerForPrediction"
10
- },
11
- "hidden_act": "silu",
12
- "hidden_size": 1024,
13
- "initializer_range": 0.02,
14
- "input_token_len": 96,
15
- "intermediate_size": 2048,
16
- "max_position_embeddings": 10000,
17
- "model_type": "timer",
18
- "num_attention_heads": 8,
19
- "num_hidden_layers": 8,
20
- "output_token_lens": [
21
- 96
22
- ],
23
- "rope_theta": 10000,
24
- "torch_dtype": "float32",
25
- "transformers_version": "4.40.1",
26
- "use_cache": true
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTm1/epoch-1/generation_config.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "transformers_version": "4.40.1"
4
- }
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTm1/epoch-1/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:44d7235dc6683442cae34df82713fc797a5d8dd02b41881720a6596fd0ca10c4
3
- size 336580760
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTm2/epoch-1/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "thuml/timer-base-84m",
3
- "architectures": [
4
- "TimerForPrediction"
5
- ],
6
- "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "thuml/timer-base-84m--configuration_timer.TimerConfig",
9
- "AutoModelForCausalLM": "thuml/timer-base-84m--modeling_timer.TimerForPrediction"
10
- },
11
- "hidden_act": "silu",
12
- "hidden_size": 1024,
13
- "initializer_range": 0.02,
14
- "input_token_len": 96,
15
- "intermediate_size": 2048,
16
- "max_position_embeddings": 10000,
17
- "model_type": "timer",
18
- "num_attention_heads": 8,
19
- "num_hidden_layers": 8,
20
- "output_token_lens": [
21
- 96
22
- ],
23
- "rope_theta": 10000,
24
- "torch_dtype": "float32",
25
- "transformers_version": "4.40.1",
26
- "use_cache": true
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTm2/epoch-1/generation_config.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "transformers_version": "4.40.1"
4
- }
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/ETTm2/epoch-1/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c15032f5bfc95a00bf5df35bf655e71321c97a923420e5523a8e4e42737aa563
3
- size 336580760
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/epoch-1/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "thuml/timer-base-84m",
3
- "architectures": [
4
- "TimerForPrediction"
5
- ],
6
- "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "thuml/timer-base-84m--configuration_timer.TimerConfig",
9
- "AutoModelForCausalLM": "thuml/timer-base-84m--modeling_timer.TimerForPrediction"
10
- },
11
- "hidden_act": "silu",
12
- "hidden_size": 1024,
13
- "initializer_range": 0.02,
14
- "input_token_len": 96,
15
- "intermediate_size": 2048,
16
- "max_position_embeddings": 10000,
17
- "model_type": "timer",
18
- "num_attention_heads": 8,
19
- "num_hidden_layers": 8,
20
- "output_token_lens": [
21
- 96
22
- ],
23
- "rope_theta": 10000,
24
- "torch_dtype": "float32",
25
- "transformers_version": "4.40.1",
26
- "use_cache": true
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/epoch-1/generation_config.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "transformers_version": "4.40.1"
4
- }
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/epoch-1/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9016120917bb75ebccbdbb5f2f93bbfc7a2b8851da05294135bb42a3c087d8d
3
- size 336580760
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/epoch-2/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "thuml/timer-base-84m",
3
- "architectures": [
4
- "TimerForPrediction"
5
- ],
6
- "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "thuml/timer-base-84m--configuration_timer.TimerConfig",
9
- "AutoModelForCausalLM": "thuml/timer-base-84m--modeling_timer.TimerForPrediction"
10
- },
11
- "hidden_act": "silu",
12
- "hidden_size": 1024,
13
- "initializer_range": 0.02,
14
- "input_token_len": 96,
15
- "intermediate_size": 2048,
16
- "max_position_embeddings": 10000,
17
- "model_type": "timer",
18
- "num_attention_heads": 8,
19
- "num_hidden_layers": 8,
20
- "output_token_lens": [
21
- 96
22
- ],
23
- "rope_theta": 10000,
24
- "torch_dtype": "float32",
25
- "transformers_version": "4.40.1",
26
- "use_cache": true
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/epoch-2/generation_config.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "transformers_version": "4.40.1"
4
- }
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/epoch-2/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:01c0613a1fcd0ccb2fb5e89c4b745b9cf8a9126e0c044c41089ae968911703df
3
- size 336580760
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/epoch-3/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "thuml/timer-base-84m",
3
- "architectures": [
4
- "TimerForPrediction"
5
- ],
6
- "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "thuml/timer-base-84m--configuration_timer.TimerConfig",
9
- "AutoModelForCausalLM": "thuml/timer-base-84m--modeling_timer.TimerForPrediction"
10
- },
11
- "hidden_act": "silu",
12
- "hidden_size": 1024,
13
- "initializer_range": 0.02,
14
- "input_token_len": 96,
15
- "intermediate_size": 2048,
16
- "max_position_embeddings": 10000,
17
- "model_type": "timer",
18
- "num_attention_heads": 8,
19
- "num_hidden_layers": 8,
20
- "output_token_lens": [
21
- 96
22
- ],
23
- "rope_theta": 10000,
24
- "torch_dtype": "float32",
25
- "transformers_version": "4.40.1",
26
- "use_cache": true
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/epoch-3/generation_config.json DELETED
@@ -1,4 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "transformers_version": "4.40.1"
4
- }
 
 
 
 
 
Time_MoE/checkpoints/timerxl_finetuned/epoch-3/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b30bf8b49d3fba5a5bfc6d64ea2077115ef4adaec88906d957cca6e0c7cab428
3
- size 336580760