Update README.md
Browse files
README.md
CHANGED
|
@@ -117,7 +117,7 @@ model_name = "IQuest/IQuest-Coder-V1-40B-Thinking"
|
|
| 117 |
|
| 118 |
### Deployment with vLLM
|
| 119 |
|
| 120 |
-
For production deployment, you can use vLLM to create an OpenAI-compatible API endpoint. Please refer to the vLLM PR for implementation details.
|
| 121 |
|
| 122 |
```bash
|
| 123 |
vllm serve IQuest/IQuest-Coder-V1-40B-Instruct --tensor-parallel-size 8
|
|
|
|
| 117 |
|
| 118 |
### Deployment with vLLM
|
| 119 |
|
| 120 |
+
For production deployment, you can use vLLM to create an OpenAI-compatible API endpoint. Please refer to the [vLLM PR](https://github.com/vllm-project/vllm/pull/31575/files) for implementation details.
|
| 121 |
|
| 122 |
```bash
|
| 123 |
vllm serve IQuest/IQuest-Coder-V1-40B-Instruct --tensor-parallel-size 8
|