Update README.md
Browse files
README.md
CHANGED
|
@@ -41,7 +41,7 @@ python3 quantize_quark.py --model_dir $MODEL_DIR \
|
|
| 41 |
--exclude_layers "lm_head" \
|
| 42 |
--multi_device \
|
| 43 |
--model_export hf_format \
|
| 44 |
-
--output_dir amd/DeepSeek-R1-MXFP4
|
| 45 |
```
|
| 46 |
|
| 47 |
# Deployment
|
|
@@ -96,7 +96,7 @@ The result of AIME24 was obtained using [SGLang](https://docs.sglang.ai/) while
|
|
| 96 |
```
|
| 97 |
# Launching server
|
| 98 |
python3 -m sglang.launch_server \
|
| 99 |
-
--model /
|
| 100 |
--tp 8 \
|
| 101 |
--trust-remote-code \
|
| 102 |
--n-share-experts-fusion 8 \
|
|
|
|
| 41 |
--exclude_layers "lm_head" \
|
| 42 |
--multi_device \
|
| 43 |
--model_export hf_format \
|
| 44 |
+
--output_dir amd/DeepSeek-R1-MXFP4-Preview
|
| 45 |
```
|
| 46 |
|
| 47 |
# Deployment
|
|
|
|
| 96 |
```
|
| 97 |
# Launching server
|
| 98 |
python3 -m sglang.launch_server \
|
| 99 |
+
--model amd/DeepSeek-R1-MXFP4-Preview \
|
| 100 |
--tp 8 \
|
| 101 |
--trust-remote-code \
|
| 102 |
--n-share-experts-fusion 8 \
|