.gitattributes CHANGED
@@ -33,3 +33,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ hgpt/humanml3d-x/lora/llama-3.1-cot_maskinput_pkeep/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ hgpt/humanml3d-x/lora/llama-3.1-nocot_maskinput_pkeep/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ assets/FRoM-W1-HACT.png filter=lfs diff=lfs merge=lfs -text
39
+ assets/FRoM-W1-HGPT.png filter=lfs diff=lfs merge=lfs -text
40
+ assets/FRoM-W1-Overview.png filter=lfs diff=lfs merge=lfs -text
41
+ assets/FRoM-W1-Teaser.png filter=lfs diff=lfs merge=lfs -text
hgpt/motionx/lora/llama-3.1-cot/adapter_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lora_alpha": 16,
3
+ "r": 32,
4
+ "target_modules": [
5
+ "q_proj",
6
+ "k_proj",
7
+ "v_proj",
8
+ "o_proj",
9
+ "gate_proj",
10
+ "up_proj",
11
+ "down_proj"
12
+ ],
13
+ "modules_to_save": [
14
+ "embed_tokens",
15
+ "lm_head"
16
+ ],
17
+ "peft_type": "LORA",
18
+ "task_type": "CAUSAL_LM",
19
+ "lora_dropout": 0.05,
20
+ "bias": "none"
21
+ }
hgpt/motionx/lora/llama-3.1-cot/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc564575f8fbc594bb3aac3064c382532ab5234bd810007761d9607b4316c606
3
+ size 2302784256
hgpt/motionx/lora/llama-3.1-cot/special_tokens_map.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|end_of_text|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ }
16
+ }
hgpt/motionx/lora/llama-3.1-cot/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8636a6e43c7fcbd40a00130ca75d6177c337d841ddb312363d14fffbd577a6e
3
+ size 17307482
hgpt/motionx/lora/llama-3.1-cot/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
hgpt/motionx/lora/llama-3.1-nocot/adapter_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lora_alpha": 16,
3
+ "r": 32,
4
+ "target_modules": [
5
+ "q_proj",
6
+ "k_proj",
7
+ "v_proj",
8
+ "o_proj",
9
+ "gate_proj",
10
+ "up_proj",
11
+ "down_proj"
12
+ ],
13
+ "modules_to_save": [
14
+ "embed_tokens",
15
+ "lm_head"
16
+ ],
17
+ "peft_type": "LORA",
18
+ "task_type": "CAUSAL_LM",
19
+ "lora_dropout": 0.05,
20
+ "bias": "none"
21
+ }
hgpt/motionx/lora/llama-3.1-nocot/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd53c49b51d7744fcdd2e1ef9b422d705df095afe251c8156e0a03b4c0fa14c5
3
+ size 2302784256
hgpt/motionx/lora/llama-3.1-nocot/special_tokens_map.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|end_of_text|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ }
16
+ }
hgpt/motionx/lora/llama-3.1-nocot/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8636a6e43c7fcbd40a00130ca75d6177c337d841ddb312363d14fffbd577a6e
3
+ size 17307482
hgpt/motionx/lora/llama-3.1-nocot/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff