Update README.md
Browse files
README.md
CHANGED
|
@@ -44,7 +44,7 @@ python3 quantize_quark.py --model_dir $MODEL_DIR \
|
|
| 44 |
--skip_evaluation \
|
| 45 |
--multi_gpu \
|
| 46 |
--model_export hf_format \
|
| 47 |
-
--output_dir amd/DeepSeek-R1-0528-MXFP4
|
| 48 |
```
|
| 49 |
|
| 50 |
# Deployment
|
|
@@ -63,7 +63,7 @@ The model was evaluated on AIME24, GPQA Diamond, and MATH-500 benchmarks using t
|
|
| 63 |
</td>
|
| 64 |
<td><strong>DeepSeek-R1-0528 </strong>
|
| 65 |
</td>
|
| 66 |
-
<td><strong>DeepSeek-R1-0528-MXFP4
|
| 67 |
</td>
|
| 68 |
<td><strong>Recovery</strong>
|
| 69 |
</td>
|
|
@@ -110,14 +110,14 @@ The results of AIME24, MATH-500, and GPQA Diamond, were obtained using forked [l
|
|
| 110 |
export VLLM_QUARK_F4F6_OFFLINE_DEQUANT_TMPENVVAR=1
|
| 111 |
|
| 112 |
# Set model args
|
| 113 |
-
OUTPUT_DIR="results/DeepSeek-R1-0528-MXFP4-
|
| 114 |
LOG="logs/deepseek_0528_maxfp4.log"
|
| 115 |
|
| 116 |
# Evaluating 10 rounds
|
| 117 |
for i in $(seq 1 10); do
|
| 118 |
# seed in [0, 2**30 - 1]
|
| 119 |
SEED=$(shuf -i 0-1073741823 -n 1)
|
| 120 |
-
MODEL_ARGS="model_name=amd/DeepSeek-R1-0528-MXFP4
|
| 121 |
|
| 122 |
lighteval vllm $MODEL_ARGS "custom|aime24_single|0|0,custom|math_500_single|0|0,custom|gpqa:diamond_single|0|0" \
|
| 123 |
--use-chat-template \
|
|
|
|
| 44 |
--skip_evaluation \
|
| 45 |
--multi_gpu \
|
| 46 |
--model_export hf_format \
|
| 47 |
+
--output_dir amd/DeepSeek-R1-0528-MXFP4
|
| 48 |
```
|
| 49 |
|
| 50 |
# Deployment
|
|
|
|
| 63 |
</td>
|
| 64 |
<td><strong>DeepSeek-R1-0528 </strong>
|
| 65 |
</td>
|
| 66 |
+
<td><strong>DeepSeek-R1-0528-MXFP4 (this model)</strong>
|
| 67 |
</td>
|
| 68 |
<td><strong>Recovery</strong>
|
| 69 |
</td>
|
|
|
|
| 110 |
export VLLM_QUARK_F4F6_OFFLINE_DEQUANT_TMPENVVAR=1
|
| 111 |
|
| 112 |
# Set model args
|
| 113 |
+
OUTPUT_DIR="results/DeepSeek-R1-0528-MXFP4-Seed"
|
| 114 |
LOG="logs/deepseek_0528_maxfp4.log"
|
| 115 |
|
| 116 |
# Evaluating 10 rounds
|
| 117 |
for i in $(seq 1 10); do
|
| 118 |
# seed in [0, 2**30 - 1]
|
| 119 |
SEED=$(shuf -i 0-1073741823 -n 1)
|
| 120 |
+
MODEL_ARGS="model_name=amd/DeepSeek-R1-0528-MXFP4,dtype=bfloat16,tensor_parallel_size=8,max_model_length=71536,max_num_batched_tokens=32768,gpu_memory_utilization=0.85,generation_parameters={max_new_tokens:65536,temperature:0.6,top_p:0.95,seed:$SEED}"
|
| 121 |
|
| 122 |
lighteval vllm $MODEL_ARGS "custom|aime24_single|0|0,custom|math_500_single|0|0,custom|gpqa:diamond_single|0|0" \
|
| 123 |
--use-chat-template \
|