Upload Qwen3ForCausalLM
Browse files- README.md +6 -6
- config.json +1 -1
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +2 -2
README.md
CHANGED
|
@@ -2,13 +2,13 @@
|
|
| 2 |
license: apache-2.0
|
| 3 |
base_model: Qwen/Qwen3-4B-Instruct-2507
|
| 4 |
tags:
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
language:
|
| 10 |
-
|
| 11 |
-
|
| 12 |
---
|
| 13 |
|
| 14 |
# PoseFit Correction Model
|
|
|
|
| 2 |
license: apache-2.0
|
| 3 |
base_model: Qwen/Qwen3-4B-Instruct-2507
|
| 4 |
tags:
|
| 5 |
+
- fitness
|
| 6 |
+
- pose-correction
|
| 7 |
+
- exercise-coaching
|
| 8 |
+
- fine-tuned
|
| 9 |
language:
|
| 10 |
+
- en
|
| 11 |
+
- hu
|
| 12 |
---
|
| 13 |
|
| 14 |
# PoseFit Correction Model
|
config.json
CHANGED
|
@@ -5,7 +5,7 @@
|
|
| 5 |
"attention_bias": false,
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"bos_token_id": 151643,
|
| 8 |
-
"dtype": "
|
| 9 |
"eos_token_id": 151645,
|
| 10 |
"head_dim": 128,
|
| 11 |
"hidden_act": "silu",
|
|
|
|
| 5 |
"attention_bias": false,
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"bos_token_id": 151643,
|
| 8 |
+
"dtype": "float16",
|
| 9 |
"eos_token_id": 151645,
|
| 10 |
"head_dim": 128,
|
| 11 |
"hidden_act": "silu",
|
model-00001-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:67bea7ea3b675200c48812fe87cfd6f37dc8e83097b313bc41bd634e6a0d75e5
|
| 3 |
+
size 4967215128
|
model-00002-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e254437134f64e26e5f91c4f4a6913dfc67f0f83e5e91c7cdf67edd12e0f0d6e
|
| 3 |
+
size 3077766464
|